Defensive Measures

This section outlines the key components that have been identified for the Defensive Measures dimension as well as maturity levels within the components. These components and levels form the basis for assessing how threat informed an organization’s Defensive program is. This assessment can be conducted using the companion spreadsheet published with this white paper.

Foundational Security 1

The degree to which threat informs and prioritizes preventative security measures.

  1. None

  2. Ad Hoc patching, limited asset inventory, basic security measures

  3. Several mitigations and security controls 2 connected to relevant threats implemented, key attack surfaces and critical assets identified

  4. Knowledge of threat informs a risk management process to prioritize a set of mitigations and controls

  5. Prioritized 3 automated patching 4, attack surfaces understood, full asset inventory mapped to business operations and threats, hygiene best-practices implemented

Data Collection

Is the right data being collected based on the needs identified from analysis of threat intelligence?

  1. None

  2. Minimal visibility (e.g., single network sensor at network boundary)

  3. Compliant with best practices for network and devices (e.g., logs collected from each device according to the manufacturer’s recommendations)

  4. Threat-informed detection requirements guide sensor configuration and deployment 5 (e.g., additional Sysmon configuration driven by detection needs for ATT&CK Techniques)

  5. Threat-Optimized (Sensors evaluated, configured, and deployed to meet all threat-informed detection needs)

Detection Engineering

How much are detection analytics designed, tested, and tuned to optimize precision, recall, and robustness for relevant malicious behaviors?

  1. None

  2. Import rules / analytics from open repository

  3. Prioritize and tune imported rules / analytics from repository

  4. Testing and tuning of custom detection analytics

  5. Detection analytics developed based on knowledge of low-variance behaviors, customized to reduce false positives while maintaining robust 6 recall 7

Incident Response

How automated, strategic, and effective are responsive measures against top-priority threats?

  1. None

  2. Ad Hoc, Manual, Reactive

  3. Playbook-enabled, partially automated

  4. Informed by knowledge of threat actor (e.g., initial detection leads to follow-on investigation to detect other malicious actions expected in the campaign based on CTI) Proactive hunts are conducted driven by threat information rather than only alerts from existing analytics.

  5. Strategic, holistic, optimized to deter future events (e.g., with an understanding of the full campaign and the adversary’s likely reaction to defensive response, the defenders take decisive and coordinated actions that effectively evict the adversary such that it is not easy for them to return)

Deception Operations 8

How extensive and effective are deception operations to enable defensive objectives and the collection of new threat intelligence?

  1. None

  2. Sandboxing of suspicious executables (e.g., email attachment detonation before delivery)

  3. 1 to several Honey* (pot, token, document…) deployed and monitored, enabling detection of malicious use and early warning

  4. Honey network deployed and monitored

  5. Intentional, long-term deception operations in a realistic honey network

References

1

https://d3fend.mitre.org/

2

https://mitre-engenuity.org/cybersecurity/center-for-threat-informed-defense/our-work/nist-800-53-control-mappings/

3

https://www.first.org/epss/

4

https://mitre-engenuity.org/cybersecurity/center-for-threat-informed-defense/our-work/mapping-attck-to-cve-for-impact/

5

https://mitre-engenuity.org/cybersecurity/center-for-threat-informed-defense/our-work/atomic-data-sources/

6

https://center-for-threat-informed-defense.github.io/summiting-the-pyramid/

7

https://center-for-threat-informed-defense.github.io/summiting-the-pyramid/definitions/

8

https://engage.mitre.org/