Skip to content

reg1reg1/Metrics

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

33 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Cyber Metrics: Abstract

Quantifying cyber risk is one of the biggest challenges in info sec today. Since the advent of the Orange book, people have been trying to quantify security or measure it in some form. Does the phrase "less" or "more secure" have any inference ? Though security may never be directly measured, can there be some metrics for it. There are various frameworks, and organizations such as PCI-DSS, NIST standards, BSIMM which focuses on software security.

This endeavour of cyber metrics done by the research team at NYU asks the question whether a reliable metrics of cyber metrics can be formed by asking the gut feeling of cybersecurity experts.

NYU's cyber metrics index of expert opinion or "gut feeling" on status of cyber security has been collecting opinion data over the last 8 years. This is an attempt to parse through and analyze the data with the ELK stack and conclude whether any corelations can be drawn between "gut feeling of experts" and "the actual state of cyber security".

Overview

A part of a team of 6 students in the Fall semester, working to find relation between various aspects of the data collected by survey started 8 years back by Dan Geer and Mukul Pareek. This repository stores the work done by me as part of contributing to the index and improving the data collection process and co-relation

Data Collection

The collection of the data has been done over the past few years by sending out survey questions to handpicked cyber security veterans and experts in fields on a monthly basis, and storing the data in a csv format. Over-time this posed challenges, as the insights that can be drawn from flat files is limited, as well as the searchability of data is also restricted. It was then collectively decided to move to better way of collecting, aggregating and visualizing this data. The collective plan is to now ship the data in a streamlined basis into a manageable format. This repository documents work done in this area using the ELK stack (Elasticsearch, Logstash , Kibana) and also outlines the plan for the future setup of the data pipeline.

The data collection involves collecting data from 2 major sources.

The one is where the expert opinion was taken from experts and the one which has been fed into the Index for the past years using survey.

The other is the source of public breach or vulnerability data which can be used to draw some sort of co-relation or implication , positive or negative and conclude whether the expert opinion in cybersecurity has a form of tangible and substantial implication in affirmation or in negation.

This is the aspect where I have been involved in this project in working , and finding patterns in the existing data as well as between existing and breach data which can lead to something conclusive.

Collection infrastructure

When dealing with data feeds which is expected to grow in size, it is important to systemize the collection and parsing of this data, so it can be scaled , pushed off to cloud if need for either arises.

Based off of the above statement, I came up with a single-pipeline based architecture for maintaining the cyber index in the years to come .

The technologies I used for this implementation are as follows:

  • Painless : Scripting Language to modify data on the fly during visual analysis on data. Helps in reduce time of reparsing and processing data in case there are minor changes required.
  • Logstash: Primarily used as a centralized "stasher". It has the remarkable simplistic execution of stitching together multiple inputs and outptuts together and reads from a configuration files. Logstash in this project was used to pull in , and segregate data from different sources including SQL, Online API , CSV and JSON.
  • Elasticsearch : Distributed and resilient noSQL data storage that enables fast querying and full-text search on data using a simple REST API and most importantly compatible with the visualization stack Kibana which can be used to discover incisive visual insights onto data.
  • Kibana: Kibana is a visual analysis tool that can be used to dig deep into the existing static and live data with a plethora of analysis tools which are easy to use and can perform sophisticated co-relations with multiple feeds of data.

For co-relation we would be trying to normalize and de-normalize data and compare them with the charts we see in the publicly available cyber data. The architectural overview is shown by the diagram below.

alt test

Challenges of Finding Co-relation Data

To be able to corelate public opinion with breach is a challenging task. The challenge of finding such a data source is three-fold.

Firstly, Finding data sources or metrics that track every breach is a daunting tasks. Many organizations do not want to make this sort of data public and there are some organizations which keep track or maintain this kind of data actively. There are some resources which summarize the data over and there is then a considerable loss of granularity which is required for co-relation.

Secondly, these are the some of the sources that I found feasible for co-relation attempts with surveyed data along with their sources. Even found, there should be some form of justified co-relation with the publicly available data.

There are some data sources which have the right amount of data , but are difficult to scrap, parse or convert due to their format and hence cannot be effectively used for co-relation.

Some of the feeds I monitored for this are as under. These may be useful in future analysis and work on the index.

https://www.hackmageddon.com/category/security/cyber-attacks-statistics/

https://www2.deloitte.com/content/dam/Deloitte/kz/Documents/risk/KZ_Deloitte_Information_Security_Survey_2014_EN.pdf

https://www.informationisbeautiful.net/visualizations/worlds-biggest-data-breaches-hacks/

The last link is the one which I chose for the further data analysis. The data shows the major data breach event that crossed 30,000 records and categorizes them on the basis of It is chosen primarily because of 3 reasons.

  1. The data has relevant granularity covers a valid timespan about breaches with respect to the survey data
  2. The data is easy to scrap and parse and compare.
  3. The expert and human opinion often tends to be influenced by the media and the major breaches, and I felt this source of data would be a good point to draw co-relations with the survey data.

Patterns in existing Data

Before, the correlation was done, I spent considerable amount of time trying to find insights in the existing survey data. These include looking for relation between different attributes or demographic or threat actors of the data. Some interesting patterns and co-relations have been highlighted below. The main aim is to find a reason or implication for the ~5000 time increase of the indexes judgement of Things getting worse.

There could be actors in the data which are causing a weighted shift or bias in the results and if so, attempt was made to isolate these biases.

Pattern-1: Mathematical Average and Serial Difference of Average

This was done to identify patterns which were not revealed by the normal unmodified data. The interesting thing is when an average is taken the pattern changes from strictly increasing to a pattern of alternate crests and peaks. The blue represents the average total score, while the green represents the serial difference of the average scores. However since the question asked is appended by "as compared to last month" , every positive response adds up (cumulatively) while the average shows the scores (isolated per year) falls and rises .

avgvsserialdiff

Pattern-2: 5 year summary of people who surveyed as things got better or worse when considering actors

Actors have 5 major split by types which are as follows.

  • Insider threat
  • Nation States
  • Strategic Rivals
  • Activists/hacktivists
  • Criminals

Over the years, the count of total survey of actor input scores which added up to negative (indicating things got better) and those which added up to positive scores are compared. They are almost equal in proportion, but the small difference has cumulated over period of months and may be the reason of the perceived aggravated increase in the opinion of things becoming worse.

Pie

Pattern-3: Year-wise split of people who said things got worse vs people who said things got better.

This is emphasizing the results of the last pie chart. The comparison of count of people who said things got worse vs people who said things got better. It is evident that the large increase is because yearly a pattern split weighs in the favour of people who said things got worse. The ratio across the years represent a ratio of count of people saying better to those who said things became worse overall which, is ~ to 1:2.In simple terms for every person who said things got better, 2 people said things got worse. This pattern seems to manifest itself across years . The survey sums up the responses of individuals for the score, and because of the ratio, the score seems to be in favour of things becoming worse cumulatively increasing over the many months of surveys.

yearwisepeople

Pattern-4: Factors contributing most

The factors which have the highest sum when aggregated (options which people said became worst) is revealed to be "targets" consisting of entities such as (is the probability of target being attacked is more or less compared to last month)

  1. autonomous network connected devices
  2. counterparties
  3. mobile devices
  4. web facing applications
  5. internet exposed devices
  6. public infra/cloud
  7. end point

contributors

Modifying Input Weight/Normalizing

Basing on the fact, that most people's natural instinct in the survey is to go for the worse or static, some of the options were given more weight in terms of influencing the outcome of the survey. However, despite multiple denormalizations, the survey outcome did not reveal anything substantial . The denormalization. The other thing is the count of people voting for a particular option does not change with weight change, and altering the output of survey to look for patterns cannot be proven to be more conclusive because certain patterns were observed.

Co-relation analysis

This is the phase where I compared the survey data with the breach data containing records of events. In terms of breaches, the amount of data being exposed has maintained an inconsistent pattern. There is not a steady increase in data breaches (setting aside the sensitivity of exposure, and industry affected)

Breach events statistics between 2004-2019

breachstats

The peaks were observed in 2016,2017 and 2018 in terms of documented records being lost, but no representation of the steady increase as the survey results have us believe. The last 5 years have seen an increase in the total no of breaches, and that can explain part of the survey response leaning in on things becoming worse.

Note that since the questions asked are appended with "as compared to last month" , so the score does not represent the current perceived state but rather the perceived increase or decrease

breachesbytype

Breach Share(in terms of sum of records) by industry or sector affected

Most of the breaches that occur are in websites or non-government based organization breaches such as yahoo. In both the no of breaches and the share of major breaches web is the dominant one. One hypothesis is -When websites are hacked , public opinion seems to be influenced as people are concerned when their private data is affected directly. The share is represented below. The first shows the share by count , and the second pie chart shows the share by no of records lost.

breachpie

breachpiecount

Data Sensitivity

The data feed used for co-relation also has the segregation of different data sources according to sensitivity. 1 being the least sensitive data and 5 being the most sensitive data. When we compare these results , and plot the pie chart for understanding the major factors contributing to breaches

The sensitivity share in the sum of all records lost over the past 6 years have been shown below.

sensitivity

A hypothesis which I felt could be applied was that "When more sensitive data gets breached, the expert opinion will be more reflective of that". The more the sensitive the breached data, it is more likely that a hardened system or government body might have leaked the data which might worry the experts and sway the opinion towards feeling things getting worse.

The graph below shows the average sum of records being split by whether they have a sensitivity of more than or less than 3.

sensitivitysplitby3

When the average of total score by year is plotted against average records lost in that year with high sensitivity(more than 3), we see that it mimics it to an extent till 2018 which had an extremely high no of records being lost. The no of high sensitivity records breached in 2018 was 2.653B. It is the point in data where the experts did not reflect it in their opinion. However till 2017 the expert average score seems to mimic the breach data pattern.

averagebreachvsorig

Improving the inference strength

During the my work in the metrics project I felt that there are some calculations we must make before we can enforce some of the inferences drawn from the data as being tangible or useful to a greater degree. These are as follows:

According to csis.org, the total cybersecurity workforce is around 700k. Assuming this as the population size, the current sample size of 44, the confidence in the results would be 95% with a huge expense in margin of error as 15%.

With better surveying methods, the sample size needs to be increased to strengthen the inferences.

The following is the number of people who filled the survey per year. The average response size over the past few years is at an even lower 30.4

surveysamplesize

Conclusions

There are some minor patterns and mimics that can be seen as highlighted in the graphs above, but the conclusions can only be reinforced with a better sample size of response. It is clear in some aspects that things have gotten worse (no of average records ) being breached per year and hence the expert opinion seems to show a trend in that direction. There are some analysis which might be revealed in future data, and where the data storage is designed from the get-go to feed the analysis part of it.

About

Cyber Metrics Elastic Indexer

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages