This section has two subsections dedicated to presenting information regarding the synthesis and evaluation of information gathered from the sources identified in the previous section.
4.1 Information synthesis
Throughout the review process, we maintained our database in the form of a spreadsheet that comprises the following sheets:
-
Search highlights sheet that contains the search strings, search engine, date of search, overall hits as well as hits after applying the inclusion and exclusion criteria.
-
Highlights sheet with the following headings: title of study, name of authors, publication date, summary, conclusion, recommendations, properties of cyber security metrics. This is followed by the proposed cyber security metrics.
Cyber security metrics sheet with the following headings: study reference number, publication date and type. This is followed by the proposed cyber security metric along with its description. This is followed by category and subcategory. The categories and subcategories are adopted from [1] (see Table 2). In addition to these categories, we have included one more category named other, whereby we list any metrics that do not fall under the main four categories. Additionally, to enhance the categorisation of metrics, we extended the subcategories of all categories to include additional ones that were elicited from some of the selected sources.
Table 2
Summary of the main categories and subcategories
Category
|
Subcategories
|
Remarks
|
Vulnerability
|
User, Interface-induced, and Software
|
pertains to measuring the level of system vulnerability
|
Defence
|
Preventive, Reactive, and Proactive
|
aims to measure the strength of cyber defence
|
Attack
|
Zero-day, Targeted, Botnet, Malware, and Evasion techniques
|
aims to measure the strength of cyber attacks
|
Situation
|
Security state, Incidents, and Investment
|
reflects the development of attack-defence interaction
|
In addition to these categories, we have included one more category named other, whereby we list any metrics that do not fall under the main four categories. Additionally, to enhance the categorisation of metrics, we extended the subcategories to include additional ones that were elicited from the following sources: 1, 3, 4, 8, 9, 18, 21, 24, and 26. They are as follows: configuration management, access control management, backup and restore, security audit, security testing, and security training. In addition to probability-based, time-based, ideal-based, and design-based metrics.
-
SMART assessment sheet lists all the identified cyber security metrics in the previous stage in addition to five columns that represent the SMART criteria so that we could assess each metric accordingly.
-
Properties of cyber security metrics sheet contains the key attributes that a metric should possess.
This approach allowed us to gather and keep track of all the relevant information in a way that enabled us not just to get insights into the direction and trends of the research in this area, but also to find gaps in the existing metrics, their properties and classification.
4.2 Information analysis of SMART metrics
In this subsection, we discuss the approach we followed to analyse the information gathered in the previous subsection. In terms of source types, 70 percent of the selected sources are conference proceedings, 13 percent of which used case study as their method, whereas four percent of the journal articles are survey research papers.
We analysed the selected sources by studying their contribution to security metrics research in two key aspects, one of which is proposal of metrics and/or proposal of properties that a good metric should have. Moreover, when gaps and/or unanswered questions found, we added them under the problems heading of the sheet. In addition to recommendations for future research directions.
The main three headings of the cyber security metrics sheet are the metrics, categories, and subcategories, which are the main elements of this work. Most of the sources proposed security metrics. Nonetheless, the ones that did not propose any metrics explicitly they either addressed the definition or properties of metrics. We recorded all the proposed metrics accompanied by their proposed description whenever provided, although some sources did not include a description with their proposed metrics and, hence, we noted them with no description.
Moreover, the authors in [33] recommend that “metrics should be SMART (Specific, Measurable, Actionable, Relevant, and Time-dependent)”. According to [13], the SMART criteria has been used widely for developing good security metrics. Thus, we assessed all the identified metrics against the SMART criteria. Each metric needs to satisfy the five elements of the criteria. Therefore, the total number of security metrics identified from 24 sources are 100, all of which fall under five categories and 26 subcategories. While the total number of the security metrics after applying the SMART criteria is 22, all of which fall under four categories and 10 subcategories.
The statistics of security metrics count (C) can be broken down further, to count how many times a metric is identified (i.e., per source) per category/subcategory as outlined in Table 3.
Table 3
Count of sources from which security metrics were identified per category/subcategory
Category
|
Subcategory
|
Count
|
System Vulnerabilities
|
Interface-Induced Vulnerabilities
|
13
|
System Vulnerabilities
|
Software Vulnerabilities
|
8
|
System Vulnerabilities
|
Overall Vulnerabilities
|
5
|
System Vulnerabilities
|
User Vulnerabilities
|
5
|
Situation
|
Security Incidents
|
9
|
Situation
|
Security State
|
5
|
Situation
|
Cost
|
4
|
Situation
|
Resilience Metrics
|
3
|
Situation
|
Agility Metrics
|
1
|
Other
|
Backup and Restore
|
4
|
Other
|
Access Control Management
|
3
|
Other
|
Configuration Management
|
3
|
Other
|
Security Training
|
3
|
Other
|
Probability-Based Metrics
|
2
|
Other
|
Compliance
|
1
|
Similarly, Table 4 shows the counts of sources per category after applying the SMART criteria.
Table 4
Count of sources from which SMART security metrics were identified
Category
|
Subcategory
|
Count
|
Sources
|
System Vulnerabilities
|
Software Vulnerabilities
|
6
|
2, 3, 5, 8, 19, 25, 30
|
System Vulnerabilities
|
Interface-Induced Vulnerabilities
|
5
|
3, 9, 25 (2), 27
|
System Vulnerabilities
|
Overall Vulnerabilities
|
4
|
4, 9, 15, 35
|
System Vulnerabilities
|
User Vulnerabilities
|
5
|
2, 3, 8 (2), 18
|
Other
|
Configuration Management
|
3
|
3, 8, 29
|
Other
|
Security Training
|
3
|
9, 15, 18
|
Defence Power
|
Intrusion Detection
|
4
|
3, 15, 19, 29
|
Defence Power
|
Proactive Defences Strength
|
2
|
3, 20
|
Defence Power
|
Preventative Defences Strength
|
1
|
19
|
Situation
|
Security Incidents
|
6
|
5 (2), 8, 9, 30, 31
|
Moreover, we view the security metrics per category or subcategory. Hence, to present the data visually, we first imported the two datasets (total cyber security metrics and the SMART security metrics) to Elasticsearch - “the world’s leading free and open search and analytics solution” [34], and then we used Kibana to create various visualisations, such as data table, tag cloud, pie chart, etc. For instance, we started creating visualisations for the first dataset to show statistics in relation to the percentage of metrics that fall under each category and subcategory so that we can gain insight into where most metrics are categorised. Additionally, we created visualisations to show security metrics per category. This is useful when we look at the metrics from the categorisation perspective. Therefore, we began by creating a tag cloud visualisation to show all the identified metrics as shown in Fig. 1.
We note that vulnerability count metric is the top metric followed by attack surface and mean time between incidents.
Moreover, some visualisations could not show all data due to the size constraint that yield in missing labels from pie charts; this happens when there are many slices (i.e., it is not possible to visualise all the labels in one figure). To solve this problem, we created a visualisation for each category so that we could show all the security metrics and subcategories within each category and then add it to the main dashboard. However, this was not necessary for the attack/threat severity category because there is a total of three security metrics identified all of which did not satisfy the SMART criteria. Hence, since the focus of this work is on the assessed security metrics, we present the visualisations that address the SMART security metrics.
We start with the category named other, Fig. 2 shows a pie chart of the assessed security metrics (i.e., after applying the SMART criteria) that fall under this category accompanied by the percentage of how many times a metric was identified out of the total number of the identified metrics within the category. To do this, under slice by, we add the category field first, and then add the subcategory field as a second slice, aggregated by the security metric field. Hence, the inner circle represents the category, the middle circle shows the subcategories, and the outer circle shows the security metrics.
We see that the top security metric in this category is the trained personnel count, identified in three sources. The authors in [9] describe this metric as “percentage of information system security personnel that have received security training”.
In Fig. 3, we present the security metrics that fall under the situation category.
The top security metric in this category is the unauthorised access attempts, identified in three sources. The authors in [5] describe this metric as “high number of failed authentication attempts”.
In Fig. 4, we present the security metrics that fall under the defence power category.
The top security metric in this category is the detection performance, identified in four sources. The authors in [3] describe this metric as “a measure of the effectiveness of the detection mechanisms (intrusion detection system, anti-virus software, etc.)”.
In Fig. 5, we present the security metrics that fall under the system vulnerabilities category.
The top security metric in this category is the vulnerability count, identified in nine sources. The authors in [1] describe this metric as the number of systems that haven’t been patched yet. The authors add that it normally takes a while for all the required patches to be applied successfully.
We created one more visualisation that shows all the SMART security metrics as shown in Fig. 6.
We can see that majority of the identified metrics fall under system vulnerability category with 51 percent, 18 percent of which fall under software vulnerability subcategory, while 13 percent fall under interface-induced vulnerability and 11 and 9 percent fall under user and overall vulnerability subcategory respectively. Hence, the top security metrics is the vulnerability count with 16 percent. Similarly, the second top category is defence power with 20 percent that comprises three subcategories intrusion detection 13 percent, proactive defences strength four percent, and preventative defences strength 2 percent. Hence, the top security metric is detection performance with nine percent.
In Table 5, we summarise the security metrics that satisfied the SMART criteria. This is followed by a column that contains the category and subcategory the security metrics fall under. This is followed by a column that contains the number of times a metric was identified from the selected sources.
Table 5
Count of sources from which SMART security metrics were identified
Metric
|
Category
|
Subcategory
|
Count
|
Sources
|
Vulnerability Count
|
System Vulnerabilities
|
Software Vulnerabilities
|
6
|
2, 3, 5, 8, 19, 30
|
Vulnerability Count
|
System Vulnerabilities
|
Overall Vulnerabilities
|
3
|
4, 15, 35
|
Detection Performance
|
Defence Power
|
Intrusion Detection
|
4
|
3, 15, 19, 29
|
Password Guessability
|
System Vulnerabilities
|
User Vulnerabilities
|
4
|
2, 3, 8, 18
|
Access Points Count
|
System Vulnerabilities
|
Interface-Induced Vulnerabilities
|
3
|
3, 9, 25
|
Trained Personnel Count
|
Other
|
Security Training
|
3
|
9, 15, 18
|
Unauthorised Access Attempts
|
Situation
|
Security Incidents
|
3
|
5, 8, 30
|
Detection Mechanism Deficiency Count
|
Defence Power
|
Intrusion Detection
|
2
|
3, 15
|
Patch Count
|
Other
|
Configuration Management
|
2
|
8, 29
|
Access Retries Count
|
System Vulnerabilities
|
Interface-Induced Vulnerabilities
|
1
|
31
|
Administrative Privileges Users Count
|
Other
|
Configuration Management
|
1
|
3
|
Critical Network Nodes
|
System Vulnerabilities
|
Interface-Induced Vulnerabilities
|
1
|
25
|
Defence in Depth Layers Count
|
Defence Power
|
Proactive Defences Strength
|
1
|
3
|
Events Assignment Delay
|
Situation
|
Security Incidents
|
1
|
31
|
Malware Detector Effectiveness
|
Defence Power
|
Proactive Defences Strength
|
1
|
20
|
Non-authorised Devices Count
|
System Vulnerabilities
|
User Vulnerabilities
|
1
|
8
|
Percentage of Severe Systems
|
System Vulnerabilities
|
Interface-Induced Vulnerabilities
|
1
|
27
|
Physical Incidents Percentage
|
Situation
|
Security Incidents
|
1
|
9
|
Prevention Performance
|
Defence Power
|
Preventative Defences Strength
|
1
|
19
|
Reported Incidents Percentage
|
Situation
|
Security Incidents
|
1
|
9
|
Reported Security Holes
|
System Vulnerabilities
|
Software Vulnerabilities
|
1
|
15
|
Simultaneous Logins Count
|
Situation
|
Security Incidents
|
1
|
5
|
Vulnerabilities Mitigation Count
|
System Vulnerabilities
|
Overall Vulnerabilities
|
1
|
9
|
We note that vulnerability count metric is categorised under two subcategories software and overall respectively. This is due to the different description given by the authors from which we identified the metrics. For instance, the authors in [3] describe vulnerability count as “the sum of known and unpatched vulnerabilities, each multiplied by their exposure time interval” which falls under software vulnerability subcategory. While the authors in [35] describe vulnerability count as “measures the number of vulnerabilities present in a system” which falls under overall vulnerability subcategory. That is why it is important to develop an appropriate definition or description for the desired metrics right from the start.
4.3 Properties of good metrics
In this subsection, we discuss some of the attributes that a good metric should possess.
According to [17], security metrics are classified according to their properties. The author discusses that security metrics need to be objective and verifiable. The author [29] argues that top level security metrics should be dynamic. The authors in [32] recommend that a good metric should be repeatable and verifiable. The authors in [15] discuss that in addition to repeatable, metrics should be feasible, quantifiable, and objective. The authors in [1] agree that metrics should be quantifiable, and they add that they should also be easy to understand. The authors in [12] argue that a good metric needs to be efficient and cost effective.
Although, the authors in [9] argue that cost effective characteristic is difficult to achieve. According to [18], in addition to meaningfulness, objectiveness and cost effectiveness, it is important that we should be able to collect metrics by automatic means. The authors in [13] recommend that metrics should be comparable and reproducible. In addition to reproducible, the authors in [22] recommend that metrics should be time-bound. The authors in [9] argue that security metrics should incorporate time.
Moreover, we identified all the relevant properties from the selected sources and then compared them to each element of the SMART criteria (Specific, Measurable, Actionable, Relevant, and Timely). In this way, on one hand, we would verify the elements of the SMART criteria, and on the other hand we would be able to enhance the criteria by extending it to include properties identified in the literature that can add value in relation to assessing security metrics.
Table 6 summarises the identified properties accompanied by the number of times a property was identified, followed by the element of the SMART criteria that can be matched with or equivalent to such a property, this is followed by some remarks to denote whether a property matches, equivalent to, or can be considered as a variant of one of the five elements of the criteria.
Table 6
Summary of the properties that a good security metric should have.
Property
|
Count
|
SMART
|
Sources
|
Remarks
|
Objective
|
7
|
S
|
3, 9, 14, 15, 17, 18, 22
|
equivalent to Specific
|
Quantifiable
|
5
|
M
|
2, 3, 13, 15, 31
|
equivalent to Measurable
|
Cost effective
|
4
|
No match
|
9, 12, 13, 18
|
inexpensive to gather
|
Repeatable
|
3
|
~M
|
13, 15, 32
|
considered as a variant of Measurable
|
Reproducible
|
3
|
~M
|
9, 13, 22
|
considered as a variant of Measurable
|
Verifiable
|
3
|
No match
|
9, 17, 32
|
independently verifiable via an outside reference
|
Meaningful
|
3
|
R
|
9, 18, 31
|
equivalent to Relevant
|
Auto collected
|
2
|
A
|
18, 31
|
i.e., it can be collected by automated means
|
Consistent
|
2
|
~T
|
14, 31
|
considered as a variant of Timely
|
Easy to understand
|
2
|
~S
|
2, 3
|
considered as a variant of Specific
|
Efficient
|
2
|
~R
|
12, 14
|
considered as a variant of Relevant
|
Comparable
|
1
|
~S
|
13
|
considered as a variant of Specific
|
Feasible
|
1
|
A
|
15
|
1) equivalent to Actionable
|
Timely
|
1
|
T
|
22
|
Same
|
We see that most of the identified properties are equivalent to an element of the SMART criteria, some of which can be considered as variants. The two unique attributes are cost effective and verifiable. Thus, the SMART criteria should be extended to include these two attributes.
Finally, in terms of any future efforts to implement a subset of the identified security metrics, there are two additional aspects that need to be looked at, one of which is the scale and the other is scope. The author [29] discusses that security metrics should have a proper scale that corresponds with the analysis type. The authors in [13] argue that several security metrics lack from an appropriate description of scale. The authors suggest that scale can be nominal, ordinal, interval, ratio, or absolute. While scope can be user, software, hardware, or organisation.