Threat Modeling Process

Author: Larry Conklin
Contributor(s): Victoria Drake, Sven strittmatter, Zoe Braiterman

Introduction

This document describes a structured approach to application threat modeling that enables you to identify, quantify, and address the security risks associated with an application.

Threat modeling looks at a system from a potential attacker’s perspective, as opposed to a defender’s viewpoint. Making threat modeling a core component of your SDLC can help increase product security.

The threat modeling process can be decomposed into three high level steps. Each step is documented as it is carried out. The resulting document is the threat model for the application.

Step 1: Decompose the Application

The first step in the threat modeling process is concerned with gaining an understanding of the application and how it interacts with external entities. This involves:

  • Creating use cases to understand how the application is used.
  • Identifying entry points to see where a potential attacker could interact with the application.
  • Identifying assets, i.e. items or areas that the attacker would be interested in.
  • Identifying trust levels that represent the access rights that the application will grant to external entities.

This information is documented in a resulting Threat Model document. It is also used to produce data flow diagrams (DFDs) for the application. The DFDs show the different paths through the system, highlighting the privilege boundaries.

Step 2: Determine and Rank Threats

Critical to the identification of threats is using a threat categorization methodology. A threat categorization such as STRIDE can be used, or the Application Security Frame (ASF) that defines threat categories such as Auditing & Logging, Authentication, Authorization, Configuration Management, Data Protection in Storage and Transit, Data Validation, and Exception Management.

The goal of the threat categorization is to help identify threats both from the attacker (STRIDE) and the defensive perspective (ASF). DFDs produced in step 1 help to identify the potential threat targets from the attacker’s perspective, such as data sources, processes, data flows, and interactions with users.

These threats can be classified further as the roots for threat trees; there is one tree for each threat goal. From the defensive perspective, ASF categorization helps to identify the threats as weaknesses of security controls for such threats. Common threat lists with examples can help in the identification of such threats. Use and abuse cases can illustrate how existing protective measures could be bypassed, or where a lack of such protection exists. The determination of the security risk for each threat can be made using a value-based risk model such as DREAD, or a less subjective qualitative risk model based upon general risk factors (e.g. likelihood and impact).

Step 3: Determine Countermeasures and Mitigation

A vulnerability may be mitigated with the implementation of a countermeasure. Such countermeasures can be identified using threat-countermeasure mapping lists. Once a risk ranking is assigned to the threats in step 2, it is possible to sort threats from the highest to the lowest risk and prioritize mitigation efforts.

The risk mitigation strategy might involve evaluating these threats from the business impact they pose. Once the possible impact is identified, options for addressing the risk include:

  • Accept: decide that the business impact is acceptable
  • Eliminate: remove components that make the vulnerability possible
  • Mitigate: add checks or controls that reduce the risk impact, or the chances of its occurrence

The following sections examine these steps in depth and provide examples of the resulting threat model in a structured format.

Decompose the Application

The goal of this step is to gain an understanding of the application and how it interacts with external entities. This goal is achieved by information gathering and documentation. The information gathering process is carried out using a clearly defined structure, which ensures the correct information is collected.

Threat Model Information

Information identifying the threat model typically includes the the following:

  1. Application Name: The name of the application examined.
  2. Application Version: The version of the application examined.
  3. Description: A high level description of the application.
  4. Document Owner: The owner of the threat modeling document.
  5. Participants: The participants involved in the threat modeling process for this application.
  6. Reviewer: The reviewer(s) of the threat model.

Example:

Threat Model Information

Application Version: 1.0

Description: The college library website is the first implementation of a website to provide librarians and library patrons (students and college staff) with online services. As this is the first implementation of the website, the functionality will be limited. There will be three users of the application:

  1. Students
  2. Staff
  3. Librarians

Staff and students will be able to log in and search for books, and staff members can request books. Librarians will be able to log in, add books, add users, and search for books.

Document Owner: David Lowry

Participants: David Rook

Reviewer: Eoin Keary

External Dependencies

External dependencies are items external to the code of the application that may pose a threat to the application. These items are typically still within the control of the organization, but possibly not within the control of the development team. The first area to consider when investigating external dependencies is the production environment and requirements.

It is useful to understand how the application is or is not intended to be run. For example, if the application is expected to be run on a server that has been hardened to the organization’s hardening standard and it is expected to sit behind a firewall, then this information should be documented in the external dependencies section.

External dependencies should be documented as follows:

  1. ID: A unique ID assigned to the external dependency.
  2. Description: A textual description of the external dependency.

Example:

External Dependencies

ID Description
1 The college library website will run on a Linux server running Apache. This server will be hardened per the college’s server hardening standard. This includes the installation of the latest operating system and application security patches.
2 The database server will be MySQL and it will run on a Linux server. This server will be hardened per the college’s server hardening standard. This will include the installation of the latest operating system and application security patches.
3 The connection between the web server and the database server will be over a private network.
4 The web server is behind a firewall and the only communication available is TLS.

Entry Points

Entry points define the interfaces through which potential attackers can interact with the application or supply it with data. In order for a potential attacker to attack an application, entry points must exist. Entry points in an application can be layered. For example, each web page in a web application may contain multiple entry points.

Entry points show where data enters the system (i.e. input fields, methods) and exit points are where it leaves the system (i.e. dynamic output, methods), respectively. Entry and exit points define a trust boundary (see Trust Levels).

Entry points should be documented as follows:

  1. ID: A unique ID assigned to the entry point. This will be used to cross-reference the entry point with any threats or vulnerabilities that are identified. In the case of layered entry points, a major.minor notation should be used.
  2. Name: A descriptive name identifying the entry point and its purpose.
  3. Description: A textual description detailing the interaction or processing that occurs at the entry point.
  4. Trust Levels: The level of access required at the entry point. These will be cross-referenced with the trust levels defined later in the document.

Example:

Entry Points

ID Name Description Trust Levels
1 HTTPS Port The college library website will be only be accessible via TLS. All pages within the college library website are layered on this entry point. (1) Anonymous Web User
(2) User with Valid Login Credentials
(3) User with Invalid Login Credentials
(4) Librarian
1.1 Library Main Page The splash page for the college library website is the entry point for all users. (1) Anonymous Web User
(2) User with Valid Login Credentials
(3) User with Invalid Login Credentials
(4) Librarian
1.2 Login Page Students, faculty members and librarians must log in to the college library website before they can carry out any of the use cases. (1) Anonymous Web User
(2) User with Login Credentials
(3) User with Invalid Login Credentials
(4) Librarian
1.2.1 Login Function The login function accepts user supplied credentials and compares them with those in the database. (2) User with Valid Login Credentials
(3) User with Invalid Login Credentials
(4) Librarian
1.3 Search Entry Page The page used to enter a search query. (2) User with Valid Login Credentials
(4) Librarian

Exit Points

Exit points might prove useful when attacking the client: for example, cross-site-scripting vulnerabilities and information disclosure vulnerabilities both require an exit point for the attack to complete.

In the case of exit points from components handling confidential data (e.g. data access components), exit points lacking security controls to protect confidentiality and integrity can lead to disclosure of such confidential information to an unauthorized user.

In many cases threats enabled by exit points are related to the threats of the corresponding entry point. In the login example, error messages returned to the user via the exit point (the log in page) might allow for entry point attacks, such as account harvesting (e.g. username not found), or SQL injection (e.g. SQL exception errors).

Assets

The system must have something that the attacker is interested in; these items or areas of interest are defined as assets.

Assets are essentially targets for attackers, i.e. they are the reason threats will exist. Assets can be both physical assets and abstract assets. For example, an asset of an application might be a list of clients and their personal information; this is a physical asset. An abstract asset might be the reputation of an organization.

Assets are documented in the threat model as follows:

  1. ID: A unique ID is assigned to identify each asset. This will be used to cross-reference the asset with any threats or vulnerabilities that are identified.
  2. Name: A descriptive name that clearly identifies the asset.
  3. Description: A textual description of what the asset is and why it needs to be protected.
  4. Trust Levels: The level of access required to access the entry point is documented here. These will be cross-referenced with the trust levels defined in the next step.

Example:

Assets

ID Name Description Trust Levels
1 Library Users and Librarian Assets relating to students, faculty members, and librarians.  
1.1 User Login Details The login credentials that a student or a faculty member will use to log into the College Library website. (2) User with Valid Login Credentials
(4) Librarian
(5) Database Server Administrator
(7) web server User Process
(8) Database Read User
(9) Database Read/Write User
1.2 Librarian Login Details The login credentials that a Librarian will use to log into the College Library website. (4) Librarian
(5) Database Server Administrator
(7) web server User Process
(8) Database Read User
(9) Database Read/Write User
1.3 Personal Data The College Library website will store personal information relating to the students, faculty members, and librarians. (4) Librarian
(5) Database Server Administrator
(6) Website Administrator
(7) web server User Process
(8) Database Read User
(9) Database Read/Write User
2 System Assets relating to the underlying system.  
2.1 Availability of College Library Website The College Library website should be available 24 hours a day and can be accessed by all students, college faculty members, and librarians. (5) Database Server Administrator
(6) Website Administrator
2.2 Ability to Execute Code as a web server User This is the ability to execute source code on the web server as a web server user. (6) Website Administrator
(7) web server User Process
2.3 Ability to Execute SQL as a Database Read User This is the ability to execute SQL select queries on the database, and thus retrieve any information stored within the College Library database. (5) Database Server Administrator
(8) Database Read User
(9) Database Read/Write User
2.4 Ability to Execute SQL as a Database Read/Write User This is the ability to execute SQL. Select, insert, and update queries on the database and thus have read and write access to any information stored within the College Library database. (5) Database Server Administrator
(9) Database Read/Write User
3 Website Assets relating to the College Library website.  
3.1 Login Session This is the login session of a user to the College Library website. This user could be a student, a member of the college faculty, or a Librarian. (2) User with Valid Login Credentials
(4) Librarian
3.2 Access to the Database Server Access to the database server allows you to administer the database, giving you full access to the database users and all data contained within the database. (5) Database Server Administrator
3.3 Ability to Create Users The ability to create users would allow an individual to create new users on the system. These could be student users, faculty member users, and librarian users. (4) Librarian
(6) Website Administrator
3.4 Access to Audit Data The audit data shows all audit-able events that occurred within the College Library application by students, staff, and librarians. (6) Website Administrator

Trust Levels

Trust levels represent the access rights that the application will grant to external entities. The trust levels are cross-referenced with the entry points and assets. This allows us to define the access rights or privileges required at each entry point, and those required to interact with each asset.

Trust levels are documented in the threat model as follows:

  1. ID: A unique number is assigned to each trust level. This is used to cross-reference the trust level with the entry points and assets.
  2. Name: A descriptive name that allows you to identify the external entities that have been granted this trust level.
  3. Description: A textual description of the trust level detailing the external entity who has been granted the trust level.

Example:

Trust Levels

ID Name Description
1 Anonymous Web User A user who has connected to the college library website but has not provided valid credentials.
2 User with Valid Login Credentials A user who has connected to the college library website and has logged in using valid login credentials.
3 User with Invalid Login Credentials A user who has connected to the college library website and is attempting to log in using invalid login credentials.
4 Librarian The librarian can create users on the library website and view their personal information.
5 Database Server Administrator The database server administrator has read and write access to the database that is used by the college library website.
6 Website Administrator The Website administrator can configure the college library website.
7 web server User Process This is the process/user that the web server executes code as and authenticates itself against the database server as.
8 Database Read User The database user account used to access the database for read access.
9 Database Read/Write User The database user account used to access the database for read and write access.

Data Flow Diagrams

All of the information collected allows us to accurately model the application through the use of Data Flow Diagrams (DFDs). The DFDs will allow us to gain a better understanding of the application by providing a visual representation of how the application processes data.

Data flows show how data flows logically through the application, end to end. They allow the identification of affected components through critical points (e.g. data entering or leaving the system, storage of data) and the flow of control through these components.

The focus of the DFDs is on how data moves through the application and what happens to the data as it moves. DFDs are hierarchical in structure, so they can be used to decompose the application into subsystems and lower-level subsystems. The high-level DFD will allow us to clarify the scope of the application being modeled. The lower level iterations will allow us to focus on the specific processes involved when processing specific data.

There are a number of symbols that are used in DFDs for threat modeling. These are described below:

Symbol Name Description
Data Flow Diagram: External Entity External Entity The external entity shape is used to represent any entity outside the application that interacts with the application via an entry point.
Data Flow Diagram: Process Process The process shape represents a task that handles data within the application. The task may process the data or perform an action based on the data.
Data Flow Diagram: Multiple Process Multiple Process The multiple process shape is used to present a collection of subprocesses. The multiple process can be broken down into its subprocesses in another DFD.
Data Flow Diagram: Data Store Data Store The data store shape is used to represent locations where data is stored. Data stores do not modify the data, they only store data.
Data Flow Diagram: Data Flow Data Flow The data flow shape represents data movement within the application. The direction of the data movement is represented by the arrow.
Data Flow Diagram: Privilege Boundary Privilege Boundary The privilege boundary (or trust boundary) shape is used to represent the change of trust levels as the data flows through the application. Boundaries show any location where the level of trust changes.

Example Diagrams

Image:Data flow1.jpg

Figure 1: Data Flow Diagram for the College Library Website.

Image:Data flow2.jpg

Figure 2: User Login Data Flow Diagram for the College Library Website.

Determine and Rank Threats

Threat Categorization

The first step in the determination of threats is adopting a threat categorization. A threat categorization provides a set of categories with corresponding examples so that threats can be systematically identified in the application in a structured and repeatable manner.

STRIDE

A threat categorization such as STRIDE is useful in the identification of threats by classifying attacker goals such as:

  • Spoofing
  • Tampering
  • Repudiation
  • Information Disclosure
  • Denial of Service
  • Elevation of Privilege

A list of generic threats classified using STRIDE is provided in the following table along with their security controls:

STRIDE Threat List

Type Description Security Control
Spoofing Threat action aimed at accessing and use of another user’s credentials, such as username and password. Authentication
Tampering Threat action intending to maliciously change or modify persistent data, such as records in a database, and the alteration of data in transit between two computers over an open network, such as the Internet. Integrity
Repudiation Threat action aimed at performing prohibited operations in a system that lacks the ability to trace the operations. Non-Repudiation
Information disclosure Threat action intending to read a file that one was not granted access to, or to read data in transit. Confidentiality
Denial of service Threat action attempting to deny access to valid users, such as by making a web server temporarily unavailable or unusable. Availability
Elevation of privilege Threat action intending to gain privileged access to resources in order to gain unauthorized access to information or to compromise a system. Authorization

Threat lists based on the STRIDE model are useful in the identification of threats with regards to the attacker goals. For example, if the threat scenario is attacking the login, would the attacker brute force the password to break the authentication? If the threat scenario is to try to elevate privileges to gain another user’s privileges, would the attacker try to perform forced browsing?

Threat Analysis

A prerequisite in the analysis of threats is the understanding of the generic definition of risk. Risk is a potential for loss determined by two factors: the likelihood, or probability, that an attack will occur; and the potential impact, or cost, of such an attack occurring. Risk is calculated as:

(Probability that threat occurs) x (Cost to organization)

From the perspective of risk management, threat modeling is a systematic and strategic approach for identifying and enumerating threats to an application environment with the objective of minimizing risk and potential impact.

Threat analysis is the identification of threats to the application, and involves the analysis of each aspect of the application’s functionality, architecture, and design. It is important to identify and classify potential weaknesses that could lead to an exploit.

From the defensive perspective, the identification of threats driven by security control categorization allows a threat analyst to focus on specific vulnerabilities. Typically, the process of threat identification involves going through iterative cycles where initially all the possible threats in the threat list that apply to each component are evaluated.

At the next iteration, threats are further analyzed by exploring the attack paths, the root causes for the threat to be exploited (e.g. vulnerabilities, depicted as orange blocks below), and the necessary mitigation controls (e.g. countermeasures, depicted as green blocks below). A threat tree as shown below is useful to perform such threat analysis.

Threat Graph

Figure 3: Threat Tree Diagram.

Once common threats, vulnerabilities, and attacks are assessed, a more focused threat analysis should take into consideration use and abuse cases. By thoroughly analyzing the use scenarios, weaknesses can be identified that could lead to the realization of a threat. Abuse cases should also be identified. These abuse cases can illustrate how existing protective measures could be bypassed, or where a lack of such protection exists.

A use and abuse case graph for authentication is shown below:

Use And Abuse Case

Figure 4: Use and Misuse Cases

The outcome of threat analysis is a determination of the types of threats posed to each component of the decomposed system. This can be documented using a threat categorization such as STRIDE or ASF, the use of threat trees to determine how the threat can be exposed by a vulnerability, and use and misuse cases to further validate the lack of any countermeasures for mitigation.

Ranking of Threats

Threats can be ranked from the perspective of risk factors. By determining the risk factor posed by the various identified threats, it is possible to create a prioritized list of threats to support a risk mitigation strategy, such as prioritizing the threats to be mitigated first. Different risk factors can be used to rank threats as High, Medium, or Low risk. In general, threat risk models use different factors to model risks such as those shown below:

Risk Model Factors

Figure 5: Ranking Risk Factors.

Subjective Model: DREAD

In the Microsoft DREAD risk assessment model, risk factorization allows the assignment of values to the different influencing factors of a threat. This provides a subjective process to rank threats. To determine the ranking of a threat, the threat analyst answers questions for each factor of risk, for example:

  • Damage: How big would the damage be if the attack succeeded?
  • Reproducibility: How easy is it to reproduce an attack?
  • Exploitability: How much time, effort, and expertise is needed to exploit the threat?
  • Affected Users: If a threat were exploited, what percentage of users would be affected?
  • Discoverability: How easy is it for an attacker to discover this threat?

A point system of numbers 1-10, representing low to high severity, is used to calculate a DREAD score that can help compare one threat to another.

Example:

Threat: Malicious user views confidential information of students, faculty members and librarians.

  1. Damage potential: Threat to reputation as well as financial and legal liability (8)
  2. Reproducibility: Fully reproducible (10)
  3. Exploitability: Require to be on the same subnet or have compromised a router (7)
  4. Affected users: Affects all users (10)
  5. Discoverability: Can be found out easily (10)

Overall DREAD score for this threat: (8+10+7+10+10) / 5 = 9

The DREAD model does not have widespread use across the industry, as its ratings are subjective. Within an organization, however, models that employ subjective ranking can help to clarify priorities when addressing risk.

Qualitative Risk Model

A generic risk model considers risk as a calculation. Recall that risk is determined by the likelihood of an attack and the impact of that attack:

(Probability that threat occurs) x (Cost to organization)

The likelihood, or probability, can be determined by the ease of exploitation. This depends on the type of threat and the system characteristics, as well as considers any countermeasures that may already be in place.

The following is a set of considerations for determining ease of exploitation:

  1. Can an attacker exploit this remotely?
  2. Does the attacker need to be authenticated?
  3. Can the exploit be automated?

The impact mainly depends on the damage potential and its extent, such as the number of components that may be affected by a threat.

Questions to help determine the damage potential are:

  1. Can an attacker completely take over and manipulate the system?
  2. Can an attacker gain administration access to the system?
  3. Can an attacker crash the system?
  4. Can the attacker obtain access to sensitive information such as secrets or PII?

Questions to help determine the number of components that are affected by a threat:

  1. How many connected data sources and systems can be impacted?
  2. How many layers into infrastructure components can the threat agent traverse?

These examples help in the calculation of the overall risk values by assigning qualitative values such as High, Medium and Low to the likelihood and impact factors. In this case, using qualitative values, rather than numeric ones like in the case of the DREAD model, help avoid the ranking becoming overly subjective.

Determine Countermeasures and Mitigation

The purpose of countermeasure identification is to determine if there is some kind of protective measure (e.g. security control, policies) that can prevent a threat from being realized. Vulnerabilities are then those threats that have no countermeasures. When threats have been categorized either with STRIDE or ASF, it is possible to find appropriate countermeasures within the given category.

Provided below is a brief and limited checklist which is by no means an exhaustive list for identifying countermeasures for specific threats.

ASF Threat & Countermeasures Examples

Threat Type Countermeasure
Authentication 1. Credentials and authentication tokens are protected with encryption in storage and transit
2. Protocols are resistant to brute force, dictionary, and replay attacks
3. Strong password policies are enforced
4. Trusted server authentication is used instead of SQL authentication
5. Passwords are stored with salted hashes
6. Password resets do not reveal password hints and valid usernames
7. Account lockouts do not result in a denial of service attack
Authorization 1. Strong ACLs are used for enforcing authorized access to resources
2. Role-based access controls are used to restrict access to specific operations
3. The system follows the principle of least privilege for user and service accounts
4. Privilege separation is correctly configured within the presentation, business and data access layers
Configuration Management 1. Least privileged processes are used and service accounts with no administration capability
2. Auditing and logging of all administration activities is enabled
3. Access to configuration files and administrator interfaces is restricted to administrators
Data Protection in Storage and Transit 1. Standard encryption algorithms and correct key sizes are being used
2. Hashed message authentication codes (HMACs) are used to protect data integrity
3. Secrets (e.g. keys, confidential data ) are cryptographically protected both in transport and in storage
4. Built-in secure storage is used for protecting keys
5. No credentials and sensitive data are sent in clear text over the wire
Data Validation / Parameter Validation 1. Data type, format, length, and range checks are enforced
2. All data sent from the client is validated
3. No security decision is based upon parameters (e.g. URL parameters) that can be manipulated
4. Input filtering via allow list validation is used
5. Output encoding is used
Error Handling and Exception Management 1. All exceptions are handled in a structured manner
2. Privileges are restored to the appropriate level in case of errors and exceptions
3. Error messages are scrubbed so that no sensitive information is revealed to the attacker
User and Session Management 1. No sensitive information is stored in clear text in the cookie
2. The contents of the authentication cookies is encrypted
3. Cookies are configured to expire
4. Sessions are resistant to replay attacks
5. Secure communication channels are used to protect authentication cookies
6. User is forced to re-authenticate when performing critical functions
7. Sessions are expired at logout
Auditing and Logging 1. Sensitive information (e.g. passwords, PII) is not logged
2. Access controls (e.g. ACLs) are enforced on log files to prevent un-authorized access
3. Integrity controls (e.g. signatures) are enforced on log files to provide non-repudiation
4. Log files provide for audit trail for sensitive operations and logging of key events
5. Auditing and logging is enabled across the tiers on multiple servers

When using STRIDE, the following threat-mitigation table can be used to identify techniques that can be employed to mitigate the threats.

STRIDE Threat & Mitigation Techniques

Threat Type Mitigation Techniques
Spoofing Identity 1. Appropriate authentication
2. Protect secret data
3. Don’t store secrets
Tampering with data 1. Appropriate authorization
2. Hashes
3. MACs
4. Digital signatures
5. Tamper resistant protocols
Repudiation 1. Digital signatures
2. Timestamps
3. Audit trails
Information Disclosure 1. Authorization
2. Privacy-enhanced protocols
3. Encryption
4. Protect secrets
5. Don’t store secrets
Denial of Service 1. Appropriate authentication
2. Appropriate authorization
3. Filtering
4. Throttling
5. Quality of service
Elevation of privilege 1. Run with least privilege

Once threats and corresponding countermeasures are identified, it is possible to derive a threat profile with the following criteria:

  1. Non mitigated threats: Threats which have no countermeasures and represent vulnerabilities that can be fully exploited and cause an impact.
  2. Partially mitigated threats: Threats partially mitigated by one or more countermeasures and can only partially be exploited to cause a limited impact.
  3. Fully mitigated threats: These threats have appropriate countermeasures in place and do not expose vulnerabilities.

Complementing Code Review

Threat modeling is not an approach to reviewing code, but it does complement the security code review process. The inclusion of threat modeling early on in the Software Development Life Cycle (SDLC) can help to ensure that applications are being developed with appropriate security threat mitigations from the very beginning. This, combined with the documentation produced as part of the threat modeling process, can give code reviewers a greater understanding of the system. This allows the reviewer to see where the entry points to the application are and the associated threats with each entry point.

When source code analysis is performed outside the SDLC, such as on existing applications, threat modeling helps to clarify the complexity of source code analysis. It promotes a depth-first approach instead of a breadth-first approach. In other words, instead of reviewing all source code with equal focus, you can prioritize the security code review of components where the threat modelling indicates higher-risk threats.