Big Data-Based Security Analytics for Protecting Virtualized Infrastructures in Cloud

Slide Note
Embed
Share

This paper proposes a novel big data-based security analytics approach to detect advanced attacks in virtualized infrastructures in cloud computing. By utilizing network logs and user application logs collected from virtual machines, attack features are extracted through graph-based event correlation and machine learning techniques to identify potential attack paths and determine the presence of attacks.


Uploaded on Nov 20, 2024 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. Download presentation by click this link. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

E N D

Presentation Transcript


  1. Big Data Based Security Analytics for Protecting Virtualized Infrastructures in Cloud ECE 693 Big Data Security

  2. Paper abstract Virtualized infrastructure in cloud computing has become an attractive target for cyberattackers to launch advanced attacks. This paper proposes a novel big data based security analytics approach to detecting advanced attacks in virtualized infrastructures. Network logs as well as user application logs that are collected periodically from the guest virtual machines (VMs) are stored in the Hadoop Distributed File System (HDFS). Then, extraction of attack features is performed through graph-based event correlation and MapReduce parser based identification of potential attack paths. Next, determination of attack presence is performed through two-step machine learning, namely logistic regression is applied to calculate attack s conditional probabilities with respect to the attributes, and belief propagation is applied to calculate the belief in existence of an attack.

  3. virtualized infrastructure A virtualized infrastructure consists of virtual machines (VMs) that rely upon the software-defined multi-instance resources of the hosting hardware. The virtual machine monitor, also called hypervisor, sustains, regulates and manages the software-defined multi-instance architecture. The ability to pool different computing resources as well as enable on- demand resource scaling has led to the widespread deployment of virtualized infrastructures as an important provisioning to cloud computing services. This also has made virtualized infrastructures become an attractive target for cyberattackers to launch attacks for illegal access.

  4. Attacks to virtualized infrastructures Exploiting the software vulnerabilities within the hypervisor source code, sophisticated attacks such as VENOM (Virtualized Environment Neglected Operations Manipulation) [1] have been performed which allow an attacker to break out of a guest VM and access the underlying hypervisor. In addition, attacks such as Heartbleed [2] and Shellshock [3] which exploit the vulnerabilities within the operating system can also be used against the virtualized infrastructure to obtain login details of the guest VMs and perform attacks ranging from privilege escalation to Distributed Denial of Service (DDoS).

  5. Existing solutions Existing security approaches to protecting virtualized infrastructures generally include two types, namely malware detection and security analytics. (1) Malware detection: It usually involves two steps, (1) monitoring hooks are placed at different points within the virtualized infrastructure, (2) a regularly-updated attack signature database is used to determine attack presence. While this allows for a real-time detection of attacks, the use of a dedicated signature database makes it vulnerable to zero-day attacks for which it has no attack signatures. (2) Security analytics: It applies analytics on the various logs which are obtained at different points within the network to determine attack presence. By leveraging the huge amounts of logs generated by various security systems (e.g., intrusion detection systems (IDS), security information and event management (SIEM), etc.), applying big data analytics will be able to detect attacks which are not discovered through signature- or rule-based detection methods. While security analytics removes the need for signature database by using event correlation to detect previously undiscovered attacks, this is often not carried out in real-time and current implementations are intrinsically non-scalable. That s why this paper was written

  6. This paper a novel big data based security analytics (BDSA) approach to protecting virtualized infrastructures against advanced attacks. By making use of the network logs as well as the user application logs collected from the guest VMs which are stored in a Hadoop Distributed File System (HDFS), the BDSA approach first extracts attack features through graph-based event correlation, a MapReduce parser based identification of potential attack paths and then ascertains attack presence through two-step machine learning, namely logistic regression and belief propagation.

  7. Malware detection (1) Malware detection (1) Malware refers to any executable which is designed to compromise the integrity of the system on which it is run. There are two prominent approaches to malware detection in cloud computing, namely (a) in-VM and outside-VM interworking approach and (b) hypervisor-assisted malware detection. (1) In-VM and outside-VM interworking detection consists of an in-VM agent running within the guest VM, and a remote scrutiny server monitoring the VM s behaviour. When a potential malware execution is detected the in-VM agent sends the suspicious executable to the scrutiny server, which then uses the signature database to verify malware presence or otherwise and then informs the in-VM agent of the results.

  8. Malware detection (2) Malware detection (2) (2) Hypervisor-assisted malware detection, on the other hand, uses the underlying hypervisor to detect malware within the guest VMs. A hypervisor-assisted malware detection scheme is designed in [6] to detect botnet activity within the guest VMs. The scheme installs a network sniffer on the hypervisor to monitor external traffic as well as inter-VM traffic. Implemented on Xen, it is able to detect the presence of the Zeus botnet on the guest VMs.

  9. Security Analytics Security Analytics Common techniques of security analytics include clustering and graph-based event correlation. Clustering for security analytics: Clustering organizes data items in an unlabeled dataset into groups based on their feature similarities. For security analytics, clustering finds a pattern which generalizes the characteristics of data items, ensuring that it is well generalized to detect unknown attacks. Examples of cluster-based classifiers include K-means clustering and k-nearest neighbors, which are used in both intrusion detection and malware detection.

  10. Graph-based event correlation While clustering determines attack presence through grouping common attack characteristics, it is limited in establishing an accurate correlation which may exist between events. This makes it difficult to accurately identify the sequences of events leading to the presence of an attack within the network, as well as the entry point of the attack. Graph-based event correlation overcomes this limitation by representing the events from the logs obtained as sequences in a graph. Given a collection of logs from different points within the network (e.g., firewall logs, web server logs, etc.), these events are correlated in a graph with the event features (e.g., timestamp, source and destination IP, etc.) represented as vertices and their correlations as edges. This enables the accurate identification of the entry point which an attack enters, as well as the sequences of events which the attack undertakes.

  11. Limitations of existing approaches Limitations of existing approaches One of the limitations of existing security approaches stems from the use of a dedicated signature database for threat detection. Security analytics removes the need for signature database by correlating events from the collected logs, but they still suffer from the post-factum (i.e., after the fact) data (thus too late!) in training for threat detection. Another limitation of existing security approaches is the centralized execution process. For instance, SINBAPT [13] runs on a single host, collecting data from various points within the network and analyzing them as a single centralized process.

  12. PROPOSED APPROACH PROPOSED APPROACH The basic idea of this proposed approach is to detect (in real-time) any malware and rookit attacks via holistic efficient use of all possible information obtained from the virtualized infrastructure, e.g., various network and user application logs. Design Principle # 1 - Unsupervised classification: The attack detection system should be able to classify potential attack presence based on the data collected from the virtualized infrastructure over time. Design Principle # 2 - Holistic prediction: The attack detection system should be able to identify potential attacks by correlating events on the data collected from multiple sources in the virtualized infrastructure.

  13. 5 Design Principles Design Principle # 3 - Real-time: The attack detection system should be able to ascertain attack presence as immediately as possible so as for the appropriate countermeasures to be taken immediately. Design Principle # 4 - Efficiency: The attack detection system should be able to detect attack presence at a high computational efficiency, i.e., with as little performance overhead as possible. Design Principle # 5 - Deployability: The attack detection system should be readily deployable in production environment with minimal change required to common production environments.

  14. big data based security analytics (BDSA) approach The BDSA approach consists of two main phases, namely (1) Extraction of attack features through graph-based event correlation and MapReduce-parser-based identification of potential attack paths, and (2) Determination of attack presence via two-step machine learning, namely logistic regression and belief propagation.

  15. Component 1: Offline training training Prior to the online detection of attacks, there is actually a system initialization, in which offline training of the logistic regression classifiers is carried out, that is, the stored features are loaded from the Cassandra database to train the logistic regression classifiers. Specifically, well-known malicious as well as benign port numbers are loaded to train a logistic regression classifier to determine if the incoming/outgoing connections are indication of an attack presence. Likewise, well-known malware and legitimate applications together with their associated ports are loaded to train a logistic regression classifier to determine if the behavior of an application running within the guest VM is indication of an attack presence. These trained logistic regression classifiers are ready for online use, upon the extraction of new attack features, to determine if the potential attack paths are indicative of attack presence.

  16. Component 2: Extraction of Attack Features In the Extraction of Attack Features phase, first, it carries out Graph- Based Event Correlation. Periodically collected from the guest VMs, network and user application logs are stored in the HDFS. By assembling the information contained in these two logs, the Correlation Graph Assembler (CGA) forms correlation graphs.

  17. Component 3: Identification of Potential Attack Paths Then, it carries out the Identification of Potential Attack Paths. A MapReduce model is used to parse the correlation graphs and identify the potential attack paths, i.e., the most frequently occurring graph paths in terms of the guest VMs IP addresses. This is based on the observation that a compromised guest VM tends to generate more traffic flows as it tries to establish communication with an attacker.

  18. Component 4: Determination of Attack Presence In the Determination of Attack Presence phase, two-step machine learning is employed, namely logistic regression and belief propagation are used. While the former is used to calculate attack s conditional probabilities with respect to individual attributes, the latter is used to calculate the belief of an attack presence given these conditional attributes. From the potential attack paths, the monitored features are sorted out and passed into their logistic regression classifiers to calculate attack s conditional probabilities with respect to individual attributes. The conditional probabilities with respect to individual attributes are passed into belief propagation to calculate the belief of attack presence. Once attack presence is ascertained, the administrator is alarmed of the attack. Furthermore, the Cassandra database is updated with the newly-identified attack features versus the class ascertained (i.e., attack or benign), which are then used to retrain the logistic regression classifiers.

  19. Extraction of Attack Features Extraction of Attack Features Use Graph-Based Event Correlation TShark is used to obtain the network logs containing the traffic flows of the guest VMs. Specifically it collects the source and the destination IP addresses along with their respective port numbers. It also undertakes the remote execution of the netstat command to obtain the guest VMs memory process lists. The network logs contain connection entries describing the guest VMs internal as well as external network connections, namely the source and destination IP addresses (i.e., IPsource and IPdestination) as well as the port numbers (i.e., Portsource and Portdestination) used.

  20. Logs Network logs: User application logs: The user application logs, on the other hand, contain process entries detailing the applications running within the guest VMs and the port numbers on which the applications are listening for connections.

  21. Form graphs Once obtained, the log entries are used to form a correlation graph based on the following observations. The first observation is that a compromised guest VM tends to communicate more frequently with other guest VMs, resulting in an increase in the network traffic containing its IP address. The second observation is that the communication methods of the malware running on the compromised VM is through its execution on it and listening for external connections. In the light of these two observations, a correlation graph is formed which best describes the guest VMs behavior by assembling the information on the network and user application logs.

  22. Graph logs The formed correlation graph, consisting of multiple paths is then stored in the HDFS on the HPC node as a new entry, called correlated log, with the format as below.

  23. MapReduce Parser for the identification of potential attack paths Identification of potential attack paths is carried out by parsing the correlation graph with a MapReduce model. MapReduce is a distributed programming model which consists of two processes namely Map and Reduce. In the Map process, key-value pairs of the form (ki, vi) are sorted from the correlation graph, where (1) ki denotes the monitored traffic flow (log format) while (2) vi is the count of occurrences of the traffic flow in the graph. Taking the correlation graph in Figure 2 (showed before) as an example, the Map process represents each path in the graph as a key ki and its occurrence as a value vi as shown in Snippet 1 (next slide).

  24. Vi

  25. Reduce phase In the Reduce process, the key-value pairs obtained during the Map process are unified. The Reduce process analyzes the intermediate key-value pairs generated from the Map process, and unifies all those key-value pairs, aggregating their occurrence counts, if their source IPs as well as source ports are the same regardless of the other elements on the path. This generates a set of new-variant key-value pairs (k0i, v0i ), where k0i represents the unified path for a distinctive source IP and port, while v0i is the total occurrence counts within the graph. For the example correlation graph above, the (unified path, aggregated counts) pairs are shown in Snippet 2 (next slide).

  26. Flagged up by the MapReduce parsers output, any graph paths with an occurrence count greater than 1 are potential attack paths and are thus picked up and passed onto the determination of attack presence phase (next slide). For the example correlation graph above, the potential attack paths are identified (marked in red) as shown in Figure 3 below.

  27. Determination of attack presence Determination of attack presence The potential attack paths identified of the correlation graph as flagged up by the MapReduce parser, can be readily retrieved into the different attack features. We refer to the stripping process as attack feature Sorter out of attack paths. For the determination of attack presence two-step machine learning is used, namely logistic regression and belief propagation. Logistic regression provides a quick means of (1) ascertaining whether a given test data projects to one of the two pre-defined classes, as well as (2) supporting the quick training of a classifier given a training set, (X Y ), which denotes a series of features versus classes. This makes it suitable for calculating attack s conditional probabilities with respect to (wrt) individual attributes. Furthermore, whenever an attack presence has been ascertained, the logistic regression classifiers can be quickly retrained in real-time using the newly- identified attack features for future attack detection.

  28. Belief propagation Belief propagation takes into account the conditional probabilities (calculated from logistic regression) in order to calculate the belief of attack presence within the virtualized environment. This allows for a holistic approach to attack detection, ensuring that the calculated belief accurately reflects the probability contributions from the individual attributes. In summary, the determination of attack presence consists of two phases, i.e., (1) Training and retraining of logistic regression classifiers (2) Attack classification using belief propagation.

  29. Training and retraining of logistic regression classifiers Used in binary classification problems, logistic regression provides a quick means of training a classifier which is used to determine if a particular test data projects to one of the two pre-defined classes. Those features will be used as conditions when calculating conditional probability in logistic regression. For each data point (x), there are n features; Total there are N data points. Y is classification result.

  30. Attributes ( deduced from features) Four attributes are defined to characterize a potential attack, namely (1) incoming network connections (in connect), (2) outgoing network connections (out connect), (3) unknown binary executions (unknown executions) and (4) opened ports (port change). While the monitored features refer to the sensor data out of the computer system being monitored, attributes are defined to characterize the situation where an attack may present. The first two attributes are used to determine attack presence based on their source and destination port numbers, while the latter two attributes are used to determine attack presence based on the applications running within the guest VMs as well as the ports opened by the applications.

  31. In order to determine the presence of the attack with respect to the attributes , logistic regression classifiers are trained for analyzing the source and destination ports as well as the applications and the ports which are opened in the guest VMs. Logistic regression calculates the probability P of attack at which a feature x maps to one of the two pre-defined classes using the logit function as below. (calculate the probability given a data point)

  32. In the context of our BDSA approach, we set two logistic regression classifiers LRapp and LRport using Eq. 4. Once trained, beforehand in a batch, and retrained with newly identified attack features, the conditional probabilities with respect to individual attributes are calculated using the respective logit functions. To train a logistic classifier for port analysis we have gathered a set of 300 port numbers used by different malware applications as well as another set of 300 ports used by legitimate applications (e.g., SSH) to be the training data set. While the malware port numbers are obtained from SANS, the port numbers used by legitimate applications are obtained from IANA (Internet Assigned Numbers Authority) In order to train the logistic regression classifier, the obtained ports are first categorized into two groups, namely system port containing the legitimate port numbers, and malware port containing the malware port numbers. Each of the port categories is then encoded with a numerical value, with system port assigned a value of 1 while malware port a value of 2, so that they can be represented as feature vectors port during the training of the logistic regression classifier.

  33. Table 1 shows the port numbers together with their encoded port category values and their classifications, with 0 representing a legitimate port and 1 representing a malware port.

  34. For Application Logs The above discussions were for Network Logs ; Similarly, to train a logistic regression classifier for application logs we have identified benign internet interfacing user applications (e.g., firefox for web browsing, nginx for web server) as well as those applications that are frequently used by malware and botnet programs (e.g., netcat). The identified applications are categorized into three categories: web app, sys util, and unknown, depending on their usages. Each of the application category is then encoded with a numerical value with web app assigned a value of 1, sys util a value of 2, and unknown a value of 3. Similarly each of the user ID is encoded with a numerical value, with user ID 0 (i.e., root user) assigned a value of 0, and user ID 1000 (i.e., non-root user) assigned a value of 1.

  35. Attack classification using belief propagation Presence of attack is determined by analyzing four attributes, namely incoming network connections (in connect), outgoing network connections (out connect), unknown binary executions (unknown exect) and opened ports (port change). This is based on the observation that the presence of an attack tends to result in changes in these attributes, as the infected guest VM attempts to establish external connections with the remote attacker. With each attribute represented by a node, they form a Bayesian network as illustrated in Figure 4(a).

  36. Belief propagation Used in graphical models such as Bayesian networks and Markov Random Fields (MRF), belief propagation is used calculate the probability distribution (i.e., belief) of a target node s state using message passing. Given a node v in a Bayesian network, the belief BEL(v) of its state is calculated using the marginal probabilities from its neighboring nodes. Belief propagation takes into account the neighboring nodes individual influence in calculating the belief of v s state, and is therefore used in our BDSA approach for determining attack presence.

  37. the port and application logistic regression classifiers (i.e., LRport and LRapp, respectively) which are trained using scikit-learn, produce (as outputs) their respective conditional probabilities, which calculate the probabilities of each feature belonging to each of the two pre-defined classes (i.e., Attack and Benign) which are of the form as below.

Related


More Related Content