7 KPIs that Make or Break an MDR PoC

  • 50caf68d5f1bc.png

    Track 7 KPIs: MTTD, MTTR(-C), Precision/Recall, Noise Index, Coverage, Response Success Rate, and Ops & Collaboration Quality (SLA/reporting).

  • Benchmarks (typical enterprise targets):

    • MTTD: minutes—< 15–30 min for top-tier

    • MTTR(-C): < 60 min to contain/respond (shorter for high‑risk)

    • Precision/Recall: >80% / >90% in PoC scenarios

    • Noise Index: operator‑digestible; e.g., < 10 alerts/day per 1,000 EP

    • Coverage: >90% key assets onboarded; >80% priority TTP coverage

    • Response Success: >90% of high‑risk incidents contained/eradicated

    • Ops Quality: SLA ≥95%, <15 min first response; IOC + timeline + clear actions in every case note



KPI 1 — MTTD (Mean Time To Detect)

What it means: Time from threat occurrence to first MDR detection/flag.

How to measure: Difference between the attack simulation timestamp (MITRE ATT&CK‑based; e.g., ransomware initial access, credential dump) and the first detection alert time.

Benchmark: Minutes to tens of minutes; top‑tier < 15–30 min.

Pro tip: Record both UTC and local timestamps; correct for data ingestion latency.



KPI 2 — MTTR‑C / MTTR (Contain / Respond)

What it means: Time from detection to containment/blocking and actionable guidance completed.

How to measure: First detection alert → endpoint isolation / account disable / IOC block (or playbook action) completed.

Benchmark: < 60 min (tighter for high‑risk); automation shortens this materially.

Pro tip: Separate acknowledge, contain, and eradicate to see where time is spent.



KPI 3 — Detection Quality (Precision & Recall)

Precision = TP / (TP + FP): share of alerts that were real → reduces waste.

Recall = TP / (TP + FN): share of real threats caught → reduces misses.

How to measure: Pre‑agreed red‑team/ATT&CK emulation scenarios; label ground truth.

Benchmark: PoC goal Precision >80%, Recall >90% (tune by environment).

Pro tip: Track per‑scenario metrics (e.g., C2 beacon vs. LOLBins) to avoid averages hiding gaps.



KPI 4 — Noise Index (False Positives & Alert Load)

What it means: Volume and ratio of “unnecessary alerts” over day/week.

How to measure: FP/day, FP%, and analyst‑actionable alerts per 100 endpoints/day.

Benchmark: Operator‑digestible; e.g., < 10/day per 1,000 endpoints.

Pro tip: Require vendors to turn on deduplication and suppression they’d recommend in production.



KPI 5 — Coverage (Visibility & ATT&CK Technique Breadth)

What it means: Which telemetry/logs are collected (EPP/EDR, IDP, email, cloud, network) and which ATT&CK TTPs are detectable.

How to measure: % asset/platform onboarding + % ATT&CK T/TA covered + data latency.

Benchmark: >90% of critical assets onboarded; >80% of priority TTPs (ransomware, credential abuse, Living‑off‑the‑Land, data exfiltration).

Pro tip: List required connectors by “Required / Optional” and verify ingest health.



KPI 6 — Response Success Rate

What it means: Share of high‑risk incidents where risk was actually reduced (contain/isolate/block and/or customer‑executed recommendation).

How to measure: Count success vs. misses across high‑risk cases; if MDR is advisory‑only, include customer action rate.

Benchmark: >90% within PoC scope.

Pro tip: Ask for per‑case evidence: IOC lists, command/process trees, network traces.



KPI 7 — Ops & Collaboration Quality (SLA, Communication, Evidence)

What it means: SLA adherence (first response, escalation), report reproducibility, handoff quality, 24×7 coverage.

How to measure: SLA attainment %, mean first‑response time, case‑note quality checklist (IOCs, timeline, clear recommendations).

Benchmark: ≥95% SLA, <15 min first response, consistent evidence in every report.

Pro tip: Standardize case templates during PoC to make vendor comparison fair.



How to Run Your PoC (Quick Checklist)

  1. Scenario set: ransomware initial access, credential dump, anomalous PowerShell, C2 beacon, O365 account takeover, exfil attempt.

  2. Measure per scenario: MTTD, MTTR, Precision/Recall, FP count, response success, report quality (1–5).

  3. Weighting example (100 pts): MTTD 20, MTTR 20, Precision/Recall 20, Noise 10, Coverage 15, Response 10, Ops 5.

  4. Normalize time: dual timestamps; correct for ingestion delay.

  5. Equal footing: same assets, same time window, same detections enabled.


Detect. Contain. Done. Breezeway’s got you covered.