Statistics for Cyber Security

statistics for cyber security n.w
1 / 27
Embed
Share

Explore the concepts of causation and association in statistics for cyber security, including correlations, linear regression, and the relationship between variables. Discover how these statistical methods help analyze data and draw meaningful insights to enhance cybersecurity measures.

  • Cyber Security
  • Statistics
  • Causation
  • Association
  • Regression

Uploaded on | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

You are allowed to download the files provided on this website for personal or commercial use, subject to the condition that they are used lawfully. All files are the property of their respective owners.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.

E N D

Presentation Transcript


  1. Statistics for Cyber Security Wenyaw Chan Department of Biostatistics School of Public Health University of Texas - Health Science Center at Houston

  2. Module (b): Correlation and Linear Regression

  3. Causation and Association Causation Changes in A cause changes in B Association: The relationship between the two variables.

  4. Causation and Association Causation After a law compelling motorists to wear seat belts went into effect, an increasing percentage of motorists complied. A study found high positive correlation between the percent of motorists wearing seat belts and the percent reduction in injuries from the day the law went into effect. This is an instance of cause and effect: Seat belts prevent injuries when an accident occurs, so an increase in their use caused a drop in injuries.

  5. Causation and Association Association A moderate correlation exists between the Scholastic Aptitude Test (SAT) scores of high school students and their grade index later as freshman in college. Surely high SAT scores do not cause high freshman grades. Rather the same combination of ability and knowledge shows itself in both high SAT scores and high grades. Both of the observed variables are responding to the same unobserved variable and this is the reason for the correlation between them.

  6. Association of Two Variables To Summarize Data taken from One Variable: Mean, Median, Mode, quantiles, Variance, Standard Deviation, Range, etc. To Summarize Data taken from Two Variables: Summary statistics of each variable and the association of the two. The association statistics include correlation, covariance, contingency table (discrete) and Regression equation (continuous).

  7. Association of Two Variables X Y Method Dichotomous Continuous Polytomous Categorical Two-sample t test Continuous One-Way ANOVA Categorical Chi-Square test or Kappa statistics Continuous Linear Regression Dichotomous Logistic Regression Continuous Continuous or Discrete

  8. Linear Regression Simple Linear Regression = 1,2,.., i n = iY + + Y x i i i ,where are independent random variables is another observable variable is the intercept is the slope is normally distributed with mean=0 and variance= ix i 2

  9. Fitting a Linear Regression Model Y S

  10. Fitting a Linear Regression Model To fit a linear regression model , we minimize the sum of squared deviations n y = 2 ( ) x i i 1 i This method is called the method of least squares.

  11. Simple Regression From the method of least squares, we can obtain the estimates of and as n ( )( ) n n x x y y = y x i i i i = = 1 i = = = 1 1 i i y x n ( ) 2 x x n i = 1 i n n = = x y i where and y i = 1 n i x = 1 n i

  12. Simple Regression The predicted value of y for a given value of x is . The properties of the Least Square estimators: 1. 2. have minimum variances among unbiased estimators (called Gauss-Markov property). 3. The residual is defined as . The sum of all the residuals is zero. 4. The regression line always goes through the point . ( ) , x y = + y x ( ) ( ) = = E and E and y y i i

  13. Linear Regression Interpretation of the Coefficients In a linear regression model, means the expected rate of increase or decrease in Y for each unit increment of x. When x increases by one unit, the mean of Y increases by units. In a linear regression model, means the expected value of Y when x=0.

  14. Simple Regression In the above equation, If , then the scatter plot of (x,y) forms a line. If >0, then as x increases, the expected value of y increases. If <0, then as x increases, the expected value of y decreases. If =0, then there is no relationship between x and y. = 2 0

  15. Simple Regression Example: Let Y be the number of computers cleaned for every 1000 times that the Malicious Software Removal Tool is run for a country and X be the country s gross income per capita. Then the regression model could be , where represents the country. i i i i Y x = + + Here, we assume that and ( ) i Var e = ( , 0, i j Cov e e for i j = ( ) = 2 Var Y i )

  16. Regression and Correlation S S = Y XY S S X XY is the sample correlation between X and Y. is the sample standard deviation of X. is the sample standard deviation of Y. X Y

  17. Some Observations of Linear Regression y 1) If we didn t have the regression line, we would use as an estimate of the yi s. 2) So is the distance our estimate is from our actual value. 3) The (directional) distance from yi to the line is This difference is called the residual component. This residual is the distance our regression estimate is from the actual variable even though we have the line. So we have improved our estimate but we still are somewhat off from the actual value. iy y y y i i

  18. Some Observations of Linear Regression 4) The distance by which we have improved our estimate for yi is . This difference is called the regression component. We have Total sum of squares = residual sum of squares + regression sum of squares. iy y

  19. An ANOVA Table for Simple Linear Regression F-Ratio=MSR/MSE df=1,n-2 for testing H0:slope=0 Source Sum of Squares Degrees of Freedom Mean Squares ( ) Model 1 MSR =SSR/1 n = 2 1 ( 1 SSR y y = i i Residual ) n-2 MSE= SSE/n-2 n = 2 y SSE y = i i i Total n-1 ( ) n 2 y y = i 1 i

  20. Extension to Multiple Linear Regression To fit a multiple linear regression model Y x x = + + + + + x 1 1 2 2 i i i k ki i we minimize the sum of squared deviations n 2 ( ) y 1 1 x 2 2 x k ki x i i i = 1 i

  21. Multiple Linear Regression Interpretation of the Coefficients In a multiple linear regression model, means the expected units of increase or decrease in Y for each unit increment of when all other x s are held constant. j j x

  22. Correlation Coefficient The correlation coefficient measures the strength of linear relationship. 1) If all we want to know is the size of the correlation coefficient, then X and Y should be continuous variables, but neither of them has to be normally distributed. 2) However, the associated hypothesis test is only valid if the pair (X,Y) are randomly selected.

  23. Correlation Coefficient ( )( ) n x x y y i i = 1 i ( ) ( i ) n n 2 2 x x y y i i = = 1 1 i

  24. Correlation Coefficient

  25. Outliers and Influential Points in a Linear Regression An outlier is an observation that appears at one of the extremes of the data range (eg: data points more than 2 sd away from the mean, . In a linear regression, an outlier can be evaluated based on three criteria: reasonableness, response extremeness and predictor extremeness (eg: a data point that has a unusually large residual). An influential point is a point which has an important influence on the coefficients of the fitted regression lines. Outliers and influential points are not necessary the same. An outlier may or may not be influential, depending on its location relative to the remaining sample points.

  26. Use of Outliers to Detect the Security Attack Outlier detection is to find patterns in data that do not conform to expected normal data behavior. Outlier detection has been a widely researched problem and find immense use in cyber security. This can be used for detecting malicious computer break-ins, called intrusion detection. Application of Outlier detection in this domain usually involves a huge volume of data.

  27. Application of Regression Model in Computer Security Y=Security breach rate. X1=number of potential adversaries X2=incentives of each adversary X3=cost of attacks X4=risk of attacks We can then run a linear regression model to predict the security breach rate using X1-X4.

More Related Content