The maximum amount of time that a system resource can remain unavailable before there is an unacceptable impact of other system resources.
RPO (Recovery Point Objective):
The point in time before a disruption or system outage to which business process data can be recovered after an outage. Given the most recent back up copy of the data
MTD (Maximum Tolerable Downtime):
The total amount of time the system owner or authorizing official is willing to accept outage or disruption. The MTD includes all impact considerations.
WRT (Work Recovery Time):
The amount of time needed to make business functions work again after the technology element is recovered. This recovery time is identified by the RTO.
Example:
RTO = 2 hrs, RPO = 10 minutes, WRT = 4 hrs, MTD = 6 hrs total to be fully functional again
Cost-Benefit Analysis of Recovery Times
Shorter RTO = Higher cost
Longer downtime = Higher operational loss
Graph (Figure 10-5): Balance between:
Cost to recover (mirror site, backups)
Cost of disruption (lost revenue, reputation)
Example: Web retail site may require real-time recovery = costly mirrored site.
Step 3 – Identify & Classify Information Assets
Classify data into Critical, Very Important, Routine, etc.
Helps prioritize what must be restored first.
Example: Payment gateway database gets higher priority than training schedules.
Step 4 – Identify Recovery Resource Requirements
List supporting assets for each business process.
Use a Resource/Component Table to document:
Hardware/software needs
Network dependencies
Staff roles
Example Table Entry:
Process: Customer Billing
Resource: Accounts Receivable Application
Details: Linux server + SQL database, ~$8,000
Step 5 – Prioritize System Resource Recovery
Use weighted scoring or simple labels (Primary/Secondary).
Avoid overcomplicating the process:
Large orgs → detailed weighted analysis.
Smaller orgs → fast classification scheme.
Output: Custom “to-do” list for disaster and continuity planning teams.
Incident Response Plan
Incident Response (IR) refers to a planned, coordinated approach to:
Detecting
Reacting to
Recovering from information security incidents.
Many organizations already engage in IR (e.g., reacting to a system crash), even if informally.
IR is essential to maintain system integrity, business continuity, and data protection.
Example: Employee accidentally deletes a key database—IT acts to restore from backup and prevent recurrence.
What is an Incident?
Adverse Event: Any unexpected event that could harm information assets.
Incident: An adverse event that materializes into a real threat.
Not every adverse event becomes an incident—but all incidents begin as adverse events.
Goal: Distinguish between routine system activity and real incidents.
This is the first phase in the NIST incident response lifecycle.
Begins with incident classification—deciding if an adverse event is an actual incident.
Sources for detection:
End-user reports
Intrusion Detection/Prevention Systems (IDPS)
Antivirus/antimalware alerts
Admin observations
Example: A help desk user notices strange pop-ups → report to CSIRT → IR plan initiated.
What Is Incident Classification?
Definition: The process of evaluating an adverse event to determine whether it qualifies as an InfoSec incident.
Requires training, clear definitions, and consistent procedures.
Classification is critical for deciding the response path (IR vs. DR/BC).
Three Types of Incident Indicators
Possible Indicators
May suggest an incident but need further investigation.
Examples:
Unfamiliar files found by users or admins
Unknown processes running in the background
Unusual system crashes or sudden reboots
Resource spikes or drops (e.g., CPU, RAM, disk space)
Tools: Windows Task Manager, UNIX/Linux resource monitors
Probable Indicators
Stronger evidence of malicious or abnormal activity.
Examples:
System activity at odd hours (e.g., midnight traffic spikes)
New user accounts with no documentation
User-reported attacks
Alerts from IDPS (though these may include false positives)
Definite Indicators
Confirmed signs that an incident is happening or has occurred.
Examples:
Dormant accounts being used unexpectedly
Log file modifications with no authorized changes
Presence of hacker or penetration tools in unauthorized locations
Notification by trusted partner or external organization
Web defacement or extortion message from a hacker
Incidence Detection Results That May Indicate an Incident
Treat all unusual results as potential incidents—better to overreact than ignore.
Possible outcomes of actual or attempted incidents:
Loss of availability (system crash or downtime)
Loss of integrity (corrupted or altered data)
Loss of confidentiality (data leak or unauthorized access)
Violation of policy (e.g., unapproved file sharing)
Violation of law/regulation (e.g., unauthorized access to protected health info)
Why Early Detection Matters
Prevents small issues from becoming large-scale incidents or disasters.
Helps the IR team:
Activate predefined IR procedures
Contain and analyze the situation quickly
Reduces:
Downtime
Reputational harm
Financial loss
From Detection to Reaction
Once an incident is confirmed and classified, the IR plan moves from Detection to Reaction.
NIST SP 800-61, Rev. 2 combines this with recovery into “Containment, Eradication, and Recovery”, while the NIST CSF separates them as “Respond” and “Recover.”
The Response Phase aims to:
Stop the incident
Minimize its impact
Prepare for recovery
Notification of Key Personnel
The CSIRT activates the alert roster to notify the appropriate individuals.
Two types of rosters:
Sequential: One person contacts everyone (accurate but slow).
Hierarchical: Each person calls others in a tree structure (faster but risk of miscommunication).
Tools: Automated systems (e.g., Preparis Portal) can streamline communication.
Example: A ransomware attack triggers automated SMS, email, and voice alerts to CSIRT and IT leadership.
Alert Messages and Communication
Alert messages include just enough information so responders can act without delay.
Alert message example: “Ransomware detected on finance server. Disconnect server and contact network admin. Follow IR SOP.”
Alert rosters must be:
Regularly updated
Tested
Rehearsed
General management, legal, HR, comms, and external partners may also need to be notified depending on the incident.
Documenting the Incident
Document:
Who did what
What happened
When actions were taken
Where the incident occurred
Why/How the event unfolded
Purpose:
Enables case studies
Aids in legal defense
Supports training and simulation
Example: Documentation helps prove compliance with due care in a breach affecting customer data.
Containment Strategies – Stopping the Attack
Identify affected systems quickly, but without full forensics analysis.
Common containment methods:
Disable compromised user accounts
Reconfigure firewalls
Shut down affected apps/services (e.g., mail server)