Otto  background

What Is the Best Vulnerability and Patch Management Process?

How to build a patch management process on the IT fundamentals that keep attackers from coming back

Connect With Us

See for yourself how policy-driven IT Automation saves time and eliminates risk.

Every incident response engagement starts with the same three questions: Do you know what's on your network? Are those systems configured and patched? Is identity and authentication locked down? If the answer to any of those is no, eviction is temporary – the attacker comes back because the fundamentals aren't there.

Attackers now exploit disclosed vulnerabilities in an average of five days, down from 32 days just two years prior (Google Threat Intelligence Group, 2024). CISA's Binding Operational Directive 22-01 maintains a Known Exploited Vulnerabilities catalog with mandatory remediation timelines. NIST SP 800-40 Rev 4 treats patching as an operational priority, not an afterthought.

How vulnerability management and patch management work together

Vulnerability scanning tells you what's broken. Patch management deploys the fix. Without vulnerability management, you're patching blind. Without patch management, your scan reports collect dust.

The relationship goes deeper than a handoff between teams. Both processes depend on the same IT foundations – the big three of cybersecurity:

  • Network inventory. If an endpoint isn't in your inventory, it doesn't get patched or scanned. A complete, real-time inventory of every endpoint, application, and service on your network is the prerequisite for both vulnerability scanning and patch deployment.

  • Configuration and patching. A known good baseline – the standard build, the approved software list, the enforced settings – is what makes anomaly detection possible. Security is the art of spotting what doesn't belong on the network, and that requires a clear picture of what normal looks like. That same baseline is what makes IT troubleshooting efficient: when a system breaks and you can compare it against a standard, you isolate the problem faster.

  • Identity and authentication. Consolidated identity management – single sign-on, centralized credential stores, least-privilege access – reduces the attack surface and simplifies monitoring. If your organization standardizes on one identity provider and everything else lives in a managed credential store, any unauthorized authentication source becomes an immediate signal. If you're a Microsoft Entra ID shop and you see Okta traffic, that's worth investigating. If your approved tools list doesn't include a service and it shows up in your logs, that's either shadow IT or something worse.

When these three foundations are in place, your security program adds a layer of expertise on top of existing IT operations rather than building from scratch. When they're missing, every remediation effort is temporary.

For a deeper breakdown of how the two disciplines differ, see Patch Management vs. Vulnerability Management. For the full vulnerability remediation lifecycle, see What Is Vulnerability Management?.

Establishing a baseline and monitoring for anomalies

The big three only work if you define what "normal" looks like and then watch for deviations. That's the baseline – and it's the shared foundation that makes both IT troubleshooting and security detection possible.

Defining your baseline

Your baseline is the documented standard that answers the three questions from the opening: what belongs on this network, how should it be configured, and who should have access?

Start with what you can enumerate: authorized endpoint types, approved operating systems and versions, sanctioned applications, expected configurations, and authorized identity providers. Then get specific. "We use Windows" is too vague to detect anything. "We run Windows 11 23H2 on HP ProBook and Dell Latitude models, with CrowdStrike and Automox agents installed, authenticating through Microsoft Entra ID" gives you a baseline where an unknown endpoint, a missing agent, or traffic to an unrecognized auth provider stands out immediately.

Don't stop at software versions. A system can be fully patched and still out of baseline – a firewall rule disabled, a management port exposed, a default credential unchanged. Document the configuration state, not just the patch level. Patching and configuration enforcement work together. One without the other leaves gaps.

How to maintain the baseline

A baseline isn't a one-time exercise. Your environment changes constantly – new endpoints, new applications, new users, configuration drift from group policy conflicts or manual overrides. Maintaining the baseline requires:

  • Automated configuration enforcement. Define the standard and let the platform enforce it. When an endpoint drifts from its security baseline – a service gets disabled, a registry key changes, a firewall rule gets overridden – the system detects and corrects it without manual intervention. Automox Worklet™ scripts handle custom configuration checks and remediations that fall outside standard patching.

  • Continuous inventory reconciliation. Compare your asset inventory against your endpoint management platform's device list on a regular cadence. New devices that appear without a corresponding onboarding record need investigation.

  • Regular access reviews. Audit user accounts, service principals, and admin credentials against your identity provider. Dormant accounts, orphaned service accounts, and privilege creep all erode the baseline over time.

Detecting anomalies in practice

Once the baseline is established, deviations become signals rather than noise. The same monitoring that helps IT catch misconfigurations and troubleshoot broken systems also catches security threats:

  • A new application appears on endpoints that shouldn't have it. That's either unapproved software or malware. Either way, it needs attention.

  • An endpoint stops reporting into the management platform. It could be offline, or it could have been tampered with to avoid detection.

  • Authentication traffic flows to an unrecognized provider. A policy violation at best, a compromised credential at worst.

  • Configuration settings revert after enforcement. Something is overriding your baseline – a conflicting group policy, a user with admin access undoing changes, or an attacker disabling security controls.

The point isn't to build a separate detection system. A well-maintained baseline turns your existing IT operations tools into an early warning system. Your configuration management flags deviations from the standard. Your inventory reconciliation catches unapproved applications. Your identity provider logs surface unauthorized access attempts. The IT work and the security work are the same work.

With the foundations in place, the next question is execution: how do you turn that baseline into a repeatable patch management process that runs at the speed attackers operate?

Building your patch management process

Step 1: Maintain a complete asset inventory

Every other step depends on this one. If 15% of your endpoints aren't in your inventory, you have a 15% blind spot – in your patching, your vulnerability scanning, and your anomaly detection.

Build and maintain a real-time inventory that covers:

  • Desktops, laptops, and servers (Windows, macOS, Linux)

  • Mobile endpoints and virtual machines

  • Third-party applications and browser extensions

  • Cloud workloads and containers

Discovery methods matter. Agent-based inventory is the most reliable for endpoints – the agent reports in regardless of network location, and a missing check-in is itself a signal. Network-based scanning covers infrastructure devices and segments where agents can't run, but misses anything off-network. Most organizations need both: agents on endpoints, network scanning for everything else, and regular reconciliation between the two to catch gaps.

Reconcile your inventory against procurement records, HR onboarding data, and your endpoint management platform's device list on a regular cadence. The delta between "what we bought" and "what's reporting in" tells you where your blind spots are – retired devices still on the network, new hires without managed endpoints, or hardware that was never enrolled.

This inventory is the starting point for the baseline described in the previous section. A cloud-native endpoint management platform gives you this visibility across on-prem, remote, and hybrid environments without requiring VPN tunnels or on-premises infrastructure.

Step 2: Define a patch prioritization policy

Not all patches carry equal weight. A feature update to a desktop productivity app doesn't demand the same urgency as a critical remote code execution (RCE) vulnerability in your VPN concentrator. Your prioritization policy should account for:

  • Severity – Common Vulnerability Scoring System (CVSS) score and vendor-assigned severity rating

  • Exploitability – Is the vulnerability listed in CISA's Known Exploited Vulnerabilities (KEV) catalog? Is there a public proof-of-concept?

  • Asset criticality – Is the affected system internet-facing, or does it hold sensitive data?

  • Patch availability – Is a vendor patch available, or do you need a compensating control?

The following priority matrix provides service-level agreement (SLA) targets aligned with industry standards and regulatory expectations:

Priority level Criteria SLA target Examples
Critical CVSS 9.0-10.0, active exploitation confirmed, or listed in CISA KEV 48 hours Zero-day RCE, actively exploited privilege escalation
High CVSS 7.0-8.9, exploit code available, internet-facing assets affected 7 days Authentication bypass, SQL injection in web application
Medium CVSS 4.0-6.9, no known exploit, limited exposure 30 days Local privilege escalation, denial-of-service in internal service
Low CVSS 0.1-3.9, informational, or feature/quality updates 90 days UI bug fix, minor third-party application update

These timelines align with CISA's Binding Operational Directive 22-01 (which requires federal agencies to remediate KEV entries within 14 days for vulnerabilities cataloged in 2021 or later) and PCI DSS v4.0.1 Requirement 6.3.3 (which requires critical security patches within 30 days of release). Adjust based on your regulatory obligations and risk tolerance.

For guidance on sorting critical security patches from routine feature updates, see how to prioritize critical patches over feature updates in the FAQ below.

Step 3: Establish a patch testing workflow

Deploying an untested patch to production is a gamble. A bad update can break line-of-business applications, corrupt configurations, or trigger blue screens across your fleet. A structured testing workflow catches these issues before they reach end users.

Stage Environment Scope Duration Success criteria
1. Lab validation Isolated test environment Patch installs correctly on representative OS builds 1-2 hours No install errors, system boots normally, patch shows as installed
2. Application compatibility Test environment with business apps Core applications launch and function after patching 2-4 hours No application crashes, key workflows complete, integrations connect
3. Pilot group deployment 5-10% of production endpoints (IT staff, volunteers) Real-world validation on diverse hardware and configurations 24-48 hours No user-reported issues, no unexpected reboots, performance baseline met
4. Staged production rollout Remaining endpoints in waves Full deployment with monitoring at each stage 1-5 days Compliance rate reaches target, no critical incidents reported
5. Post-deployment verification All patched endpoints Confirm patch installed, vulnerability remediated 24 hours Scan confirms vulnerability closed, patch inventory updated

For critical and actively exploited vulnerabilities, you may compress stages one through three into a single accelerated cycle – but never skip testing entirely. Even a 30-minute smoke test on a representative system can prevent a fleet-wide outage.

Step 4: Design your rollout strategy

Deploying patches to every endpoint simultaneously maximizes risk. If the patch causes issues, your entire organization is affected at once. Staggered rollouts limit the blast radius and give you time to catch problems before they spread.

A practical rollout strategy groups endpoints into staged deployment waves (often called "rings"). In Automox, you implement this by creating device groups with staggered policy schedules:

  • Ring 0 – IT and security team (immediate): Your own team absorbs the first impact. They're equipped to troubleshoot and report issues.

  • Ring 1 – Early adopters (24 hours after Ring 0): A cross-functional group of non-critical users who represent diverse hardware, OS versions, and application stacks.

  • Ring 2 – General population (48-72 hours after Ring 0): The bulk of your endpoints, deployed in waves by department or geography.

  • Ring 3 – Critical infrastructure (72-96 hours after Ring 0): Servers, kiosks, shared workstations, and systems with strict uptime requirements. These get patched last, during scheduled maintenance windows.

Each ring should have a defined rollback procedure. If the pilot group in Ring 1 reports application failures or performance degradation, you pause the rollout, investigate, and decide whether to proceed, modify, or roll back before moving to Ring 2.

For organizations handling monthly Microsoft security updates, this ring-based approach is especially relevant. See Best Way to Handle Patch Tuesday for a detailed walkthrough of Patch Tuesday rollout planning.

Step 5: Configure maintenance windows

Maintenance windows define when patches can be installed and when reboots are permitted. Getting this right is the difference between a smooth update cycle and a flood of helpdesk tickets from users whose machines rebooted during a presentation.

Effective maintenance windows balance three factors:

  • User disruption – Schedule installations during off-hours or low-activity periods

  • Patch urgency – Critical security patches may require shorter windows or forced reboots

  • System availability – Servers and shared workstations need windows aligned with their usage patterns

Practical guidelines for setting maintenance windows:

  • Workstations: Schedule installs overnight or during lunch hours. Allow a 4-8 hour deferral window for users to save work before a forced reboot.

  • Servers: Define a weekly or bi-weekly maintenance window (e.g., Saturday 2 am-6 am) coordinated with change management. Production servers should follow your change advisory board (CAB) process.

  • Remote endpoints: Use a cloud-native platform that patches over the internet without requiring VPN. This lets remote workers receive patches on the same schedule as on-prem users. Automox handles this natively, applying patches regardless of network location.

  • Critical/zero-day patches: Override standard windows with an emergency change process. Define in advance who can authorize an emergency patch and what the notification process looks like.

Document your maintenance windows in your patch policy and communicate them to end users. Predictability reduces resistance.

Engineering for user safety

Before you plan for nation-state actors or ransomware groups, consider what your network encounters most: normal users trying to do their jobs who occasionally make mistakes.

Least-privilege access prevents those mistakes from cascading. A user in accounting who accidentally runs a malicious attachment shouldn't be able to take down a file server. An IT admin troubleshooting a workstation shouldn't need domain admin credentials to do it. When you engineer for user safety first – limiting what any single account can do – the same controls that protect users from their own mistakes also protect the network from external threats.

This applies directly to patch management:

  • Standard user accounts reduce the risk of users disabling security settings or deferring critical patches indefinitely.

  • Tiered admin access ensures patch deployment credentials don't grant broader access than the task requires.

  • Automated enforcement removes the dependency on users doing the right thing. When patch policies apply automatically, compliance doesn't hinge on user behavior.

User safety, insider threat monitoring, and external threat defense aren't separate programs. They're layers on the same foundation.

Common barriers to effective patch management

Even with a solid process on paper, teams run into recurring obstacles that slow patching and increase exposure.

Visibility gaps

Unmanaged endpoints are unpatched endpoints. On-premises tools like Windows Server Update Services (WSUS) and System Center Configuration Manager (SCCM) can't reach endpoints outside the corporate network, and that blind spot grows with every remote hire.

Fragmented tooling

Separate tools for vulnerability scanning, OS patching, third-party updates, and configuration management mean separate consoles, separate agents, and slow handoffs between teams. Each additional tool adds cost, training overhead, and integration gaps. For an analysis of hidden costs, see The True Cost of On-Prem Patch Management.

Slow remediation cycles

The exploitation timeline covered above (five days average, 24 hours for nearly a third of exploits) leaves no margin for sluggish deployment. Yet many organizations average 97 days to patch critical vulnerabilities (Ponemon Institute).

Common causes of slow remediation:

  • Manual approval processes with multiple handoffs between security and IT

  • Testing bottlenecks from limited lab environments

  • Patch deployment tools that require VPN connectivity for remote endpoints

  • Reboot coordination challenges in 24/7 operations

Automating patch approval for pre-tested, vendor-signed updates – while reserving manual review for high-risk patches – can cut mean time to remediation from weeks to hours.

Third-party application patching

Third-party applications (browsers, PDF readers, collaboration tools) are a common attack vector that many patching programs miss. Many traditional tools handle OS updates well but leave third-party patching as a manual exercise. Your process should cover third-party applications with the same rigor as OS patches. For a deeper look at this challenge, see Third-Party Patch Management Guide.

IT and security silos

When IT operations and security teams use different tools, follow different priorities, and report to different leadership, vulnerabilities fall through the cracks. Security identifies a critical CVE, writes a ticket, and waits. IT has its own queue and patches on a different schedule.

The fix isn't a tool – it's shared ownership. Both teams need access to the same remediation dashboard, a shared mean time to remediation (MTTR) metric that neither team can game independently, and a regular standup to review open critical items. When patching is framed as an IT imperative that also serves security – rather than a security demand imposed on IT – friction drops and remediation speed goes up. IT and security should work as one integrated team, not as adversaries.

Measuring your patch management program

A patch management process without metrics is a process without accountability. Track these key performance indicators (KPIs):

  • Mean time to patch (MTTP) – Average time from patch release to deployment. Target: under 72 hours for critical, under 14 days for high.

  • Patch compliance rate – Percentage of endpoints with all applicable patches installed. Target: 95% or higher within SLA windows.

  • Patch failure rate – Percentage of deployments that fail on first attempt. A high rate signals compatibility issues or tool limitations.

  • Coverage rate – Percentage of endpoints under active patch management. Anything below 100% represents unmanaged risk.

  • Vulnerability exposure window – Time between disclosure and verified remediation. Benchmark against your SLA targets.

Report these metrics monthly to security leadership and quarterly to executive stakeholders. Trend data over time matters more than any single snapshot. For a deeper look at closing the remediation gap, see How To Reduce MTTR for Vulnerability Patching.

Start with the fundamentals

The three questions from the opening aren't just an incident response checklist. They're the measure of your program's maturity. Do you know what's on your network? Can you prove those systems are configured and patched? Is identity and authentication locked down?

Every section of this guide – the baseline, the prioritization matrix, the deployment rings, the maintenance windows – exists to make the answer "yes" and keep it that way. The organizations that sustain low MTTR over years aren't the ones with the most specialized security tooling. They're the ones that got the IT fundamentals right and built security on top of them.

Sources

Frequently asked questions

Patch Tuesday – Microsoft's monthly security update release on the second Tuesday of each month – requires a structured response plan. Best practices include classifying patches against your priority matrix within the first two hours, running lab validation same-day, deploying to a pilot group within 24 hours, and staging the full rollout across deployment rings over three to five days. For a complete playbook, see Best Way to Handle Patch Tuesday.

Start with a lab environment that mirrors your production OS builds, then test application compatibility with your core business applications. Deploy to a pilot group of 5-10% of your fleet and monitor for 24-48 hours before broader rollout. Test on representative hardware and software configurations, not just a single clean VM.

Define two approval tracks: automatic and manual. Auto-approve patches that meet low-risk criteria – vendor-signed OS security updates below CVSS 7.0 with no known compatibility issues – and route everything else through manual review with severity assessment, impact analysis, and lab testing. Automox lets you encode these rules directly into patch policies so approvals happen consistently without manual intervention.

Use a deployment ring model: Ring 0 is your IT team (immediate), Ring 1 is a cross-functional pilot group (24 hours later), Ring 2 is the general population by department or geography (48-72 hours), and Ring 3 is critical infrastructure during maintenance windows (72-96 hours). Monitor each ring for failures before advancing to the next. This limits the blast radius of a bad patch and gives you multiple checkpoints to pause or roll back.

Separate your patch policies into security and non-security tracks. Security patches follow strict SLA targets (48 hours for critical, seven days for high), while feature and quality updates follow a slower 30-90 day cadence. Automox policies let you filter by patch classification, so security updates auto-deploy while feature updates wait for manual approval.

Define windows based on endpoint type: workstations during off-hours (e.g., 10 pm-6 am) with a deferral option, servers in a weekly maintenance window governed by change management, and remote endpoints via a cloud-native tool that patches over the internet without VPN. Document each window in your patch policy and define an emergency override process for zero-day patches.

The big three – network inventory, configuration and patching, and identity and authentication – are the IT fundamentals that underpin every security program. Without knowing what's on your network, keeping those systems configured and patched, and protecting how users authenticate, specialized security tooling can't keep attackers out. These are IT imperatives that double as security imperatives. For a deeper discussion, listen to the companion podcast episode.

Vulnerability management tells you what's broken and how urgent it is. Patch management deploys the fix. In a well-built process, scan results feed directly into your patch prioritization matrix, and successful deployments close the loop by confirming the vulnerability is resolved. For a detailed comparison, see Patch Management vs. Vulnerability Management.

Dive deeper into this topic