Network Anomaly Detection Using a Commute Distance Based Approach Nguyen Lu Dang Khoa, Tahereh Babaie, and Sanjay Chawla School of Information Technologies The University of Sydney Sydney NSW 2006, Australia {khoa, tarab}@it.usyd.edu.au, [email protected]

Abstract—We propose the use of commute distance, a random walk metric, to discover anomalies in network traffic data. The commute distance based anomaly detection approach has several advantages over Principal Component Analysis (PCA), which is the method of choice for this task: (i) It generalizes both distance and density based anomaly detection techniques while PCA is primarily distance-based (ii) It is agnostic about the underlying data distribution, while PCA is based on the assumption that data follows a Gaussian distribution and (iii) It is more robust compared to PCA, i.e., a perturbation of the underlying data or changes in parameters used will have a less significant effect on the output of it than PCA. Experiments and analysis on simulated and real datasets are used to validate our claims. Keywords-network anomaly detection; principal component analysis; distance-based approach; density-based approach; commute distance based approach;

I. I NTRODUCTION Anomaly detection in the context of computer network is finding unusual and large changes in network traffic. It is an important step in order to maintain high performance and security throughout the network [1]. Anomalies can be caused by many reasons ranging from intentional attacks (e.g. Distributed Denial of Service - DDoS) to unusual network traffic (e.g. flash crowds). To be able to detect anomalies in an acceptable time ensures network problems can be fixed quickly and thus limits losses. Traditional network anomaly detection techniques use signature-based methods where a pre-defined set of patterns describing previous anomalous events is used to identify future anomalies. This method can only find known attacks and new attack patterns need to be updated over time. Therefore, techniques which can find unknown types of anomalies have been studied. In recent years, PCA has been used as a simple but effective method for network anomaly detection. However, PCA results are very sensitive to parameter settings which in turn are highly data dependent. Moreover, in some circumstances, large anomalies in turn can effect the PCA computation leading to false positives [2]. Anomaly detection has been extensively studied within the data mining community. Many techniques have been developed including those based on distance-based [3],

Zainab Zaidi Networked Systems Group, NICTA Locked Bag 9013 Alexandria, NSW 1435, Australia [email protected]

[4], [5] and density-based [6], [7] approaches. However, those anomaly detection techniques are not popular in the computer network community. In this paper, we address the problem of network anomaly detection and show the weaknesses of PCA approach. Distance and density based techniques are also applied to detect anomalies. Moreover, we propose a distance-based anomaly detection technique which uses commute distance as a metric. Commute distance is a well-known measure derived from a random walk on graph [8]. The commute distance between two nodes i and j in the graph is the number of steps that a random walk, starting from i will take to visit j and then come back to i for the first time. Unlike traditional Euclidean distance, commute distance between two nodes can capture both the distance between them and their local neighborhood densities so that we can capture both global and local anomalies using distance-based techniques [9]. The experiments show that the approach using commute distance has distinct advantages over PCA in network anomaly detection. The contributions of this paper are as follows: •

• •

We apply the commute distance metric for network anomaly detection. Commute distance is more robust to data perturbation and subsumes both distance and density based approaches. We address the weaknesses of PCA in detecting anomalies in computer network domain. We report on experiments carried out on a small-size wireless network data and a backbone network data. The results show that the commute distance based approach has a lower false positive and false negative rate in anomaly detection compared to PCA and typical distance and density based approaches.

The remainder of the paper is organized as follows. Section II reviews the network anomaly detection problem and the related work. In Section III and IV, methods to detect anomalies using PCA approach and recent data mining approaches are described. Section V presents a data mining technique to find anomalies using a commute distance measure. In Section VI, we evaluate all the approaches using experiments on simulated and real datasets. Section VII is the paper conclusion.

II. BACKGROUND A. Network Anomaly Detection Problem Since the number of attack incidents in computer networks increase dramatically, anomaly detection has been considered as necessity in all network security systems. Network anomaly detection, in the field of network security, involves finding abnormal and significant changes occurring in a backbone network traffic which correspond to novel or modified attacks. Backbone networks are main infrastructure of wide computer networks which interconnect miscellaneous networks in local environment or over wide area by providing a pathway for exchange of information between subnetworks and or LANs. Typically, a backbone network consists of access nodes called Point of Presence or PoPs, which are connected through links. Network traffic engineering uses traffic measurement matrices to carry out different tasks including load balancing, capacity planning or anomaly detection. Based on how we denote traffic source and destination in matrix, different traffic matrices can be defined at any level of granularity. In other words, by selecting the level of aggregation a particular traffic matrix can be specified. Typically, significant traffic demand in a network is known as origin-destination flows (OD flows) described as a volume of traffic flows between all pair of PoPs in a specified network. The links, where each OD flow passes through the network between source and destination, are determined in a routing matrix and consequently the superpositions of those OD flows result in the traffic observed on each links. Suppose we have a virtual network with four nodes shown in Figure 1. It is being observed in time t and there are four OD flows between the first and second node (denoted by x1,t ), first and third (denoted by x2,t ), second and third (denoted by x3,t ), and the third and fourth node (denoted by x4,t ). Therefore the link traffic measurement observed on the first link denoted by y1,t is the following superposition of passing OD flows: y1,t = x1,t + x2,t .

4 y1,t

1

x 2,t x1,t 2

Router

x 4,t 3

x3,t

Figure 1: A typical network including four nodes (links) and four OD flows in time t

Thus for all links, the equation    y1,t 1 1 0 y2,t  1 0 1  = y3,t  0 1 1 y4,t 0 0 0

is:   0 x1,t   0  x2,t  .   1 x3,t  1 x4,t

To measure over time interval (t = 1, ..., m): yt = At xt Defining time-invariant matrices yt as Y , At as A and xt as X the equation will be: Y = AX Every sudden change in an OD flow traffic X is considered to be a volume anomaly which often spans over several links in the network. Such changes can be due to a DDoS attack, a viruses spread, or even a flash crowd problem. On the other side, the number of possible OD flows in a computer network with n nodes (PoPs) is proportional to n2 . Since the number of nodes in ISP networks is considerable, highly complicated requirements and expensive facilities are needed to collect traffic measurements in the level of OD flows between access nodes. Practically most ISP networks utilize Simple Network Management Protocol (SNMP) which is a standard protocol to measure link data traffic Y in the above equation. Therefore the main problem is to infer anomalies from other data measurements. Network anomaly detection problem includes two main steps: Identification and Inference. Identification finds anomalies from link traffic measurement data while Inference involves assigning them to the OD flow anomalies. In our work, we only focus on the anomaly identification step. B. Related Work Anomaly identification can be implemented on a traditional Intrusion Detection System (IDS) or a Network Anomaly Detection System (NADS). A signature-based IDS uses a set of preconfigured and predetermined attack patterns known as signatures to catch a specific and malicious incident in network traffic which has been referred to as ‘misuse’ detection. The set of these signatures must be frequently brought up to date to recognise new emerging threats in order to reach a high-quality security performance [10]. The concept of a NADS as an alarm for peculiar system behavior was introduced by Dorothy Denning [11]. Putting together an activity profile of normal activities over an interval of time and finding the deviation from these typical behaviors, the author established a NADS approach versus a traditional IDS approach. Actually a statistical anomalybased IDS sets up a routine activity baseline based on normal network traffic assessments. Then the behavior of network traffic activity will be monitored for the action that varies from this typical activity profile. While an IDS detects a

misuse signature in network traffic, a NADS tries to identify a new or previously unknown abnormal traffic behaviour. Methodologies of network anomaly identification problem can be classified as in Figure 2.

Figure 2: Methodologies in Network Anomaly Identification Recently, dimensionality reduction approach as an influential method for unravelling unusual and abnormal patterns from data has been introduced and widely discussed in computer network community. In the first best known attempt, Lakhina et al. [12], [13], [14] proposed PCA for traffic anomaly detection. In fact PCA is a linear approach that transforms a high dimensional data space with supposed correlated variables into a new low dimensional data space with smaller number of uncorrelated components while they still capture the most variance in the data. Lakhina et al. employed the concept of PCA to divide the high-dimensional space of network traffic measurement into normal subspace including typical behavior patterns and anomalous subspace counting uncharacteristic circumstances. Networkwide anomaly detection based on PCA has appeared as an effective method for detecting a wide variety of anomalies. PCA approach has some advantages comparing with other approaches. First, this approach detects OD flow anomalies using the evaluation of correlations across single links while other approaches including [15], [16], [17], [18] utilize only single timeseries traffic from a network link which is independent of traffic on other links in a network. Second, many other methods depend on parameters adjusting and a priori knowledge of structure in traffic data while PCA represents normal and anomalous traffic behaviour directly from the observed data. However, limitations of PCA approach have also been discussed in some works. In [1], they evaluated a lot of algorithms including PCA in order to determine the strengths of methods in network-level anomaly detection from available data aggregation. As another effort in [2], the authors showed that methods for tuning PCA are not adequate and starting with a new dataset, adjusting parameters is unexpectedly difficult.

III. A N APPROACH USING PRINCIPAL COMPONENT ANALYSIS

Principal Components Analysis is one of the most common methods in dimensionality reduction. PCA transforms a high dimensional data to a linear lower-dimensional representation while the rebuilt data has as little as possible change in variance. In fact, it projects data in a multivariate space into a new subspace with smaller number of uncorrelated variables called Principle Components. The first principal component represents the variable capturing maximal variance of data and the next subsequent component corresponds to the variable capturing remaining maximal variance and is orthogonal to the first one. The next ones have smaller variances and point to the directions of remaining orthogonal principal components. Typically first few principal components account for most of the variance in the original data so that the rest of principal components can be skipped with a minimum loss of information. Mathematical calculation includes the eigenvalue decomposition of the covariance matrix of the dataset, after removing the mean of the data for each attribute. Anomalies can be detected by looking at directions defined by the first few or the last few principal components. In general, the last few principal components are more likely to contain information which does not conform to the normal data [19]. Since the first few principal components capture most of variance in the dataset, they are strongly related to one or more of the original variables. As a result, observations which are anomalies on the direction of the first few principal components are usually anomalies on one or more of the original variables as well. Anomalies in this case can be detected with statistical methods. The last few principal components represent the linear combination of original variables with very small variance. Thus the data tend to result in similar small values in those principal components. Therefore, any observation that largely deviates from this tendency for the last few components is likely to be an anomaly. There are a number of statistical tests for anomalies using PCA. Dunia and Qin [20] introduced a subspace approach based on decomposition of a main space of data into normal and anomalous subspaces using projection of the data on the first few principle components and on the last few ones respectively. In [13] Lakhina et al. proposed a network anomaly detection method based on the same subspace analysis using link traffic data. In the proposed approach, the first p eigenvectors form a subspace called normal subspace S N and the remaining m − p eigenvectors form an anomalous or residual subspace S A . Supposing a network with m data links over t time intervals, we denote link traffic matrix Xi×j = [X1 , ..., Xm ] where each row xi represents an instance of the entire link loads at time i (1 ≤ i ≤ t) and each column Xj corresponds to time series

of j-th link (1 ≤ j ≤ m). An observation xi are decomposed into two portions, A xi = xN i + xi N where xN i is a projection of xi onto S , T xN i = P P xi = Cxi

where matrix P is formed by principal components of the normal subspace S N as columns. The matrix C = P P T represents the projection onto p dimensional subspace S N . The residual xA i belongs to the m−p dimensional anomalous subspace S A , xA i = (I − C)xi The anomalies tend to result in a large change in xA i and network traffic instances are considered anomalies if d2i = 2 kxA i k is greater than a chosen threshold. IV. A NOMALY DETECTION TECHNIQUES USING RECENT DATA MINING APPROACHES

Anomaly detection has been extensively studied within statistical community [21], [22]. Statistical approach is often model-based and assumes that data follows a certain distribution. The data is evaluated by its fitness to the model generated by the assumed distribution. If the probability of a data instance (under the assumed model) is less than a threshold, it is considered an anomaly. However, the distribution assumption is often violated in practise. Knorr and Ng were the first to propose the definition of distance-based anomaly (or distance-based outlier) which does not make any assumption about the data distribution [3]. An observation o in a dataset T is a DB(p, D) anomaly if at least a fraction p of observations in T have greater distances than D from o. However, it is difficult to estimate suitable values of p and D and furthermore it is difficult to rank anomalies with these definitions. To address the problem, Ramawamay et al. [4] proposed the definition of DB(k, N ) anomaly which is based on a distance from an observation to its k-th nearest neighbor. Anomalies are the top N observations whose distances to its k-th nearest neighbor are greatest. In order to reduce the time to find the nearest neighbors, Bay and Schwabacher [5] proposed a simple but powerful pruning rule to find DB(k, N ) anomalies. An observation is not an anomaly if its current score (e.g. an average distance to its k current nearest neighbors) is less than the score of the weakest anomaly among top N anomalies found so far. Using this approach, a large number of non-anomalies can be pruned without carrying out a full data search scan. The weakness of distance-based anomaly detection techniques is they are able to detect only global anomalies and often fail to detect local anomalies in a dataset with regions

of different densities. Breunig et al. proposed the concept of density-based anomaly and its measure called Local Outlier Factor (LOF), which can identify local anomalies [6]. It captures the degree to which an observation p is an anomaly by looking at the densities of its neighbors. The lower the p’s local density and the higher the local densities of p’s nearest neighbors, the higher the LOF value of p. Observations with the largest LOF values are marked as anomalies. However, LOF is computationally expensive with the complexity of O(n2 ) [23]. Moreover, LOF could not find small anomalous clusters which are near to each other [9]. Recently, Khoa and Chawla [9] presented a new method to find anomalies using a measure called commute distance. Indeed commute distance is an Euclidean distance in the space spanned by eigenvectors of the graph Laplacian matrix. Unlike Euclidean distance, commute distance between two points can capture both the distance between them and their local neighborhood densities so that we can capture both global and local anomalies using distance-based methods such as methods in [5], [3]. Moreover, the method can be applied directly to graph data. The detail of this method is described in Section V. V. A DISTANCE - BASED APPROACH USING COMMUTE DISTANCE

A. Commute Distance Let P be the transition matrix of the random walk on graph with entry pij , A is the graph adjacency matrix, and D is the diagonal matrix with entries dii . Then P = D−1 A. The commute time, which is known to be a metric and that is the reason for the term ‘commute distance’ [24], is the expected number of steps that a random walk starting at i will take to reach j once and go back to i for the first time [8]: c(i, j) = h(i, j) + h(j, i). (1) The hitting time h(i, j) is the expected number of steps a random walk starting at i will take to reach j for the first time: ( 0 if i = j P h(i, j) = (2) 1 + k∈adj(i) pik h(k, j) otherwise where adj(i) is a set of neighbors of node i. The commute distance can be computed from the MoorePenrose pseudoinverse of the graph Laplacian matrix [25], [24]. Denote L = D − A and L+ as the graph Laplacian matrix and its pseudoinverse respectively, the commute distance is: + + + c(i, j) = VG (lii + ljj − 2lij ), (3) Pn + where VG = i=1 dii is the volume of the graph and lij is + the (i, j) element of L . Equation 3 can be written as

c(i, j) = VG (ei − ej )T L+ (ei − ej ),

(4)

where ei is the i-th column of an (n × n) identity matrix I [26]. Consequently, c(i, j)1/2 is a distance in the Euclidean space spanned by the ei ’s. B. Commute Distance Based Anomaly Detection This section introduces a method based on commute distance to detect anomalies [9]. As commute distance is a metric and is able to capture both the distance between nodes and their local neighborhood densities, commute distance based method can be used to find global and local anomalies. First, a mutual k1 -nearest neighbor graph is constructed from the dataset. The edge weights are inversely proportional to their Euclidean distances. However, it is possible that the mutual k1 -nearest neighbor graph is not connected so that we cannot apply random walk on the whole graph. One approach to make the graph connected is to find its minimum spanning tree and add the edges of the tree to the graph. Then the graph Laplacian matrix L and its pseudoinverse L+ are computed. After that the pairwise commute distances between any two observations are calculated from L+ . Finally, the distance-based anomaly detection using commute distance with pruning technique proposed by Bay and Schwabacher [5] is used to find the top N anomalies. The anomaly score used is the average commute distance of an observation to its k2 nearest neighbors. While commute distance is a robust measure for detecting both global and local anomalies, the main drawback is its computational time. The direct computation of commute distance from L+ is proportional to O(n3 ) which is not feasible for large graphs (n is the number of nodes). Khoa and Chawla proposed a graph component sampling technique and an eigenspace approximation to accelerate the computation [9]. The idea of the graph component sampling is to construct the similarity graph, sample only the normal graph components, and leave the anomalous components intact. Sampling in this way will maintain the geometry of the original graph and the relative densities of normal clusters. Anomalies are also not sampled in this sampling strategy. Eigenspace approximation transforms the commute distance in an n dimensional space to the commute distance in an m dimensional space (m ¿ n). Therefore, we just need to compute the m smallest eigenvectors with nonzero eigenvalues of L (i.e. the largest eigenvectors of L+ ) to approximate the commute distance. The approximate algorithm has a complexity of O(nlogn) [9]. VI. E XPERIMENTS AND R ESULTS A. Datasets All the methods were analysed on two sources of simulated and real datasets. The first one is from a small-sized NICTA wireless mesh network [27]. The NICTA wireless mesh network has seven nodes deployed in the School of IT at the University of Sydney. It used a traffic generator to simulate the traffic on the network. Packets were aggregated

into one-minute time bins and the data was collected in 24 hours. There were 391 OD flows and 1270 time bins. Four anomalies were introduced to the network where three of them were DOS attacks and one was a ping flood. The second one is from Abilene backbone network. It has 11 nodes and connects many universities and research labs in the United States. In Abilene, packets were aggregated into five-minute time bins. We used two weeks of Abilene data: April, 9-15th, 2004; and September, 4th-10th, 2004. Each dataset has 2016 time bins. B. Evaluation strategy For NICTA dataset, since we have labels for anomalies, the results can be evaluated easily. The problem comes from Abilene datasets which we do not have labels for anomalies. To isolate and verify anomalies in a computer network is a very difficult task. We applied a strategy mentioned in [1] to evaluate the results. The idea is using the anomalies found by a detection method directly on OD flow data as j benchmark. Specifically denoted BM as the set of top M anomalies found by the detection method j on OD flow data as benchmark, AiN as the set of top N anomalies found by the detection method i on link data. Then we can have false positives and false negatives as follows. i • False Positives: these are time bins found in AN , but j j i not in the benchmark BM , i.e., AN − BM (N < M ) • False Negatives: these are time bins found in the benchj j − AiN (N > M ) , but not in AiN , i.e., BM mark BM We chose the smaller of M and N to be 30, and the larger to be 50 like the experiment in [1]. C. Experimental Results 1) NICTA Dataset: PCA approach [13] (described in Section III) was able to detect all four anomalies in the dataset. The threshold Lakhina et. al used to choose the number of eigenvectors for normal and anomalous subspace is as follows. They examined the projection on each principal component in descending order of eigenvalues. When a projection was greater than the threshold (three time of standard deviation away from the mean), that principal component and all the subsequence principal components were assigned to the anomalous subspace. All the previous principal components belonged to the normal subspace. In this dataset, except the four anomalies, the remaining time bins locate in the direction of the first eigenvector and thus it is not difficult to detect these anomalies using PCA. Figure 3 shows the PCA plot of the dataset. However, PCA result depends largely on the number of eigenvectors for the normal subspace. In this dataset, Lakhina et al.’s technique to choose the number of eigenvectors k = 1 was successful. If k = 2, only one anomalies was found and if k = 3, all the anomalies found were incorrect. For a comparison to PCA, three different methods in data mining were used to detect anomalies. They were a

Table I: False positives in the top 30 detected anomalies compared with the top 50 benchmark anomalies

NICTA dataset Outliers 2000

Top 30 detected PCA EDOF LOF CDOF

PC3

0 −2000

False positive with top 50 benchmark PCA EDOF LOF CDOF Average 28 28 30 27 28.25 2 0 0 2 1.00 7 4 1 1 3.25 2 0 1 0 0.75

−4000 −6000 10000 5000 3000 2000

0 PC2

1000 −5000

0 −1000

PC1

Figure 3: The NICTA dataset and anomalies on the first three principal components. The variance was 0.99 for the three principal components. distance-based method [5] (denoted as EDOF), a densitybased method [6] (denoted as LOF), and a commute distance based method (described in Section V and denoted as CDOF). The threshold to choose anomalies for the three methods was three time of standard deviation away from the mean of all the anomaly scores. The 10 mutual nearest neighbor graph was used for CDOF. All three methods used k = 10 as the number of nearest neighbors to estimate the anomaly scores. The results showed EDOF and CDOF could detect all the four anomalies while LOF only captured three of them. 2) Abilene Dataset: Dataset 1 (April, 9-15th, 2004): Since labels are not available for these datasets, threshold was not used to find anomalies. We detected top N = 50 anomalies based on the anomaly scores and applied the strategy described in Section VI-B to evaluate the results. For each method (PCA, EDOF, LOF, and CDOF) we compared an anomaly set captured on link data with anomaly sets found by the direct analysis of benchmark methods on OD flow data. The number of eigenvectors for PCA was chosen using Lakhina et al.’s approach. We chose the number of nearest neighbors k = 60 for EDOF, LOF, and CDOF. Good methods should achieve the low false positives and false negatives for most of the benchmarks. Table I shows the false positives seen in the top 30 detected anomalies compared with the top 50 benchmark anomalies. It can be seen that CDOF and EDOF had the lowest average false positives over all the benchmarks and CDOF was better than EDOF. LOF had higher false positives and PCA show the highest false positives among all the methods. Table II exhibits the false negatives found in the top 50

detected anomalies compared with the top 30 benchmark anomalies. It shows that CDOF, EDOF, and LOF had relatively low average false negatives over all the benchmarks and CDOF was better than EDOF and LOF. The PCA approach almost missed all anomalies in the dataset. The bad results of PCA were probably due to the bad choice of number of eigenvectors for PCA in link flow data (k = 4 compared with k = 2 in OD flow data). Table II: False negatives in the top 50 detected anomalies compared with the top 30 benchmark anomalies Top 50 detected PCA EDOF LOF CDOF

False negative with top 30 benchmark PCA EDOF LOF CDOF Average 28 30 30 30 29.50 4 0 0 3 1.75 4 2 0 2 2.00 4 0 0 0 1.00

The next experiments show the sensitivity of parameters used for different approaches. The results are shown on Figure 4. For PCA, the range of number of eigenvectors for normal subspace k=1-10 was used. For EDOF, LOF, and CDOF, the range of number of nearest neighbors for estimating anomaly scores k=10-100 was used. For the benchmark method, k = 2 was chosen for PCA and k = 60 was chosen for EDOF, LOF, and CDOF. It is shown that PCA and LOF were more sensitive to the parameters than EDOF and CDOF in this dataset. When the number of nearest neighbors k ≥ 40, EDOF and CDOF had the best results with zero false positive and false negative and they were stable when k increased. The results for PCA and LOF varied when the parameters changed. Dataset 2 (September, 4th-10th, 2004): This experiment shows that PCA fails in case there are large anomalies affecting the normal subspace of the data. In this dataset, all EDOF, LOF, and CDOF found only one anomaly if the threshold based on anomaly scores was used while PCA could not find it. PCA returned 935 time bins as anomalies instead. An analysis on the data shows that there was a very large and dominating traffic volume appearing in time bin 1350. The reason PCA could not find it is the idea to use PCA to detect anomaly is based on the assumption that normal traffic is captured by the first few principal components and the anomalies are captured by the remaining components. However, in this case, the

30 False Positives False Negatives

# false positives/negatives

25

20

15

10

5

0

1

2

3

4 5 6 7 8 # eigenvectors for normal subspace

9

10

(a) PCA 12 False Positives False Negatives

# false positives/negatives

10

8

6

4

2

0 10

20

30

40

50 60 70 # nearest neighbors

80

90

100

(b) EDOF 25 False Positives False Negatives

# false positives/negatives

20

15

5

20

30

40

50 60 70 # nearest neighbors

80

90

100

(c) LOF 20 False Positive False Negative

18

# false positive/negative

R EFERENCES [1] Y. Zhang, Z. Ge, A. Greenberg, and M. Roughan, “Network anomography,” in IMC ’05: Proceedings of the 5th ACM SIGCOMM conference on Internet Measurement. Berkeley, CA, USA: USENIX Association, 2005, pp. 30–30. [2] H. Ringberg, A. Soule, J. Rexford, and C. Diot, “Sensitivity of pca for traffic anomaly detection,” in SIGMETRICS ’07: Proceedings of the 2007 ACM SIGMETRICS international conference on Measurement and modeling of computer systems. New York, NY, USA: ACM, 2007, pp. 109–120.

16 14 12 10 8 6 4 2 0 10

VII. C ONCLUSION In this paper we present the network intrusion detection problem and propose the use of commute distance, a random walk metric, to discover anomalies in network traffic data. We also address the weaknesses of PCA in detecting anomalies in computer network domain. PCA is very sensitive to the number of eigenvectors chosen and is incapable of detecting large anomalies appearing on the normal subspace. The experimental results on simulated and real datasets show that the commute distance based approach, which is more robust to data perturbation and is able to generalize both distance and density based anomaly detection techniques, has a lower false positive and false negative rate in anomaly detection compared to PCA and typical distance and density based approaches. ACKNOWLEDGMENT The authors of this paper acknowledge the financial support of the Capital Markets CRC.

10

0 10

an artificial anomalies as follows. A time bin 1350 was removed from the dataset. Then the principal components of the data were found. An artificial anomaly was placed on the first eigenvector and very far from all other data in the direction of the first eigenvector. It was an anomaly in term of traffic volume comparing with all the remaining data. Then we applied PCA, EDOF, LOF, and CDOF on the new dataset. EDOF, LOF, and CDOF all found the point we generated as the top anomaly. For PCA, it incorrectly classified 359 points as anomalies and none of them was the generated point.

20

30

40

50 60 70 # nearest neighbors

80

90

100

(d) CDOF

Figure 4: The sensitivity of parameters used in all approaches

large anomaly skewed the normal subspace forming by the first few principal components and consequently increased the false positives. To be more convincing on our claim above, we generated

[3] E. M. Knorr and R. T. Ng, “Algorithms for mining distancebased outliers in large datasets,” in The 24rd International Conference on Very Large Data Bases, 1998, pp. 392–403. [4] S. Ramaswamy, R. Rastogi, and K. Shim, “Efficient algorithms for mining outliers from large data sets,” in SIGMOD ’00: Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data. New York, NY, USA: ACM, 2000, pp. 427–438. [5] S. D. Bay and M. Schwabacher, “Mining distance-based outliers in near linear time with randomization and a simple pruning rule,” in KDD ’03: Proceedings of the ninth ACM SIGKDD international conference on Knowledge discovery and data mining. New York, NY, USA: ACM, 2003, pp. 29–38.

[6] M. M. Breunig, H.-P. Kriegel, R. T. Ng, and J. Sander, “Lof: Identifying density-based local outliers,” in Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, May 16-18, 2000, Dallas, Texas, USA, W. Chen, J. F. Naughton, and P. A. Bernstein, Eds. ACM, 2000, pp. 93–104.

[19] I. T. Jolliffe, Principal Component Analysis, 2nd ed. Springer, October 2002.

[7] S. Papadimitriou, H. Kitagawa, P. Gibbons, and C. Faloutsos, “Loci: fast outlier detection using the local correlation integral,” in Data Engineering, 2003. Proceedings. 19th International Conference on, March 2003, pp. 315–326.

[21] D. Hawkins, Identification of Outliers. and Hall, 1980.

[8] L. Lov´asz, “Random walks on graphs: a survey,” Combinatorics, Paul Erd¨os is Eighty, vol. 2, pp. 1–46, 1993. [9] N. L. D. Khoa and S. Chawla, “Robust outlier detection using commute time and eigenspace embedding,” in PAKDD ’10: Proceedings of the The 14th Pacific-Asia Conference on Knowledge Discovery and Data Mining. Berlin/Heidelberg: Springer, 2010, pp. 422–434. [10] H. J. M. Michael E. Whitman, Ed., Principles Of Information Security. Course Technology, 2008. [11] D. E. Denning, “An intrusion-detection model,” IEEE Trans. Softw. Eng., vol. 13, no. 2, pp. 222–232, 1987. [12] A. Lakhina, M. Crovella, and C. Diot, “Characterization of network-wide anomalies in traffic flows,” in IMC ’04: Proceedings of the 4th ACM SIGCOMM conference on Internet measurement. New York, NY, USA: ACM, 2004, pp. 201– 206. [13] ——, “Diagnosing network-wide traffic anomalies,” in SIGCOMM ’04: Proceedings of the 2004 conference on Applications, technologies, architectures, and protocols for computer communications. New York, NY, USA: ACM, 2004, pp. 219–230. [14] ——, “Mining anomalies using traffic feature distributions,” in SIGCOMM ’05: Proceedings of the 2005 conference on Applications, technologies, architectures, and protocols for computer communications. New York, NY, USA: ACM, 2005, pp. 217–228. [15] P. Barford, J. Kline, D. Plonka, and A. Ron, “A signal analysis of network traffic anomalies,” in IMW ’02: Proceedings of the 2nd ACM SIGCOMM Workshop on Internet measurment. New York, NY, USA: ACM, 2002, pp. 71–82. [16] M. Roughan, T. Griffin, M. Mao, A. Greenberg, and B. Freeman, “Combining routing and traffic data for detection of ip forwarding anomalies,” in SIGMETRICS ’04/Performance ’04: Proceedings of the joint international conference on Measurement and modeling of computer systems. New York, NY, USA: ACM, 2004, pp. 416–417. [17] B. Krishnamurthy, S. Sen, Y. Zhang, and Y. Chen, “Sketchbased change detection: methods, evaluation, and applications,” in IMC ’03: Proceedings of the 3rd ACM SIGCOMM conference on Internet measurement. New York, NY, USA: ACM, 2003, pp. 234–247. [18] J. D. Brutlag, “Aberrant behavior detection in time series for network monitoring,” in LISA ’00: Proceedings of the 14th USENIX conference on System administration. Berkeley, CA, USA: USENIX Association, 2000, pp. 139–146.

[20] R. Dunia and S. J. Qin, “Multi-dimensional fault diagnosis using a subspace approach,” in Proceedings of American Control Conference 1997, 1997, pp. 353–365. London: Chapman

[22] A. M. L. Peter J. Rousseeuw, Robust Regression and Outlier Detection. John Wiley and Sons, 2003. [23] A. B. V. Chandola and V. Kumar, “Outlier detection: A survey,” Department of Computer Science and Engineering, University of Minnesota, Twin Cities, Tech. Rep. TR 07-017, 2007. [24] F. Fouss and J.-M. Renders, “Random-walk computation of similarities between nodes of a graph with application to collaborative recommendation,” IEEE Transaction on Knowledge and Data Engineering, vol. 19, no. 3, pp. 355–369, 2007. [25] D. J. Klein and M. Randic, “Resistance distance,” Journal of Mathematical Chemistry, vol. 12, pp. 81–95, 1993. [Online]. Available: http://dx.doi.org/10.1007/BF01164627 [26] M. Saerens, F. Fouss, L. Yen, and P. Dupont, “The principal components analysis of a graph, and its relationships to spectral clustering,” in Proceedings of the 15th European Conference on Machine Learning (ECML 2004). Lecture Notes in Artificial Intelligence. Springer-Verlag, 2004, pp. 371–383. [27] Z. R. Zaidi, S. Hakami, B. Landfeldt, and T. Moors, “Real-time detection of traffic anomalies in wireless mesh networks,” Wireless Networks, 2009. [Online]. Available: http://www.springerlink.com/content/w85pp037p7614j28/

Network Anomaly Detection Using a Commute ...

Anomaly detection in the context of computer network is finding unusual and ... (e.g. Distributed Denial of Service - DDoS) to unusual network traffic (e.g. flash ...

286KB Sizes 5 Downloads 296 Views

Recommend Documents

Robust Outlier Detection Using Commute Time and ...
walk. This method cannot find outlying clusters where the node connectivities are still high. An excellent survey by Chandola et. al [4] provides a more detailed view on outlier ... Since CD is a metric, a distance based method can be used to realize

Anomaly Detection for malware identification using ...
merce or ebanking attracts criminals, who using sophisticated techniques, tries to introduce ... One of the most used approach is detecting and analyzing automated activity based on the .... After some test, we are unable to use this ... using a scri

Anomaly Detection Using Replicator Neural Networks ...
Keywords: artificial neural networks, replicator neural network, auto- encoder, anomaly ... anomaly detection does not require training data. However, it makes ...

Sparse Representation based Anomaly Detection using ...
HOMVs in given training videos. Abnormality is ... Computer vision, at current stage, provides many elegant .... Global Dictionary Training and b) Modeling usual behavior ..... for sparse coding,” in Proceedings of the 26th Annual International.

large scale anomaly detection and clustering using random walks
Sep 12, 2012 - years. And I would like to express my hearty gratitude to my wife, Le Tran, for her love, patience, and ... 1. Robust Outlier Detection Using Commute Time and Eigenspace Embedding. Nguyen Lu ..... used metric in computer science. Howev

Credit Card Fraud Detection Using Neural Network
some of the techniques used for creating false and counterfeit cards. ..... The illustration merges ... Neural network is a latest technique that is being used in.

Masquerade Detection Using IA Network
lenge to the computer security, where an illegitimate entity poses as (and assumes the identity of) a legitimate entity. The illegitimate user, called masquerader ...

Time Series Anomaly Detection - Research at Google
Statistical and regression techniques seem more promising in these cases. Netflix recently released their solution for anomaly detection in big data using Robust.

Anomaly detection techniques for a web defacement monitoring ...
DI3 – Università degli Studi di Trieste, Via Valerio 10, Trieste, Italy ... exploiting security vulnerabilities of the web hosting infrastruc-. ture, but ...... the best results.

Anomaly detection techniques for a web defacement monitoring service
jack DNS records of high profile new zealand sites, 2009; Puerto rico sites redirected in DNS attack security, 2009) ..... Hotelling's T-Square method is a test statistic measure for detecting whether an observation follows a ... statistics of the ob

A Comparative Study of Anomaly Detection Techniques ...
approach with anomaly detection techniques that have been proposed earlier for host/network-based intrusion detection systems. This study enables gaining further insights into the problem of automatic detection of web defacements. We want to ascertai

Brain Tumor Detection Using Neural Network ieee.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Brain Tumor ...

Programming Exercise 8: Anomaly Detection and ... - nicolo' marchi
multivariateGaussian.m - Computes the probability density function .... takes as input the data matrix X and should output an n-dimension vector mu that holds the ..... dients.2 If your implementation is correct, you should find that the analytical.

Sparse Representation based Anomaly Detection with ...
Sparse Representation based Anomaly. Detection with Enhanced Local Dictionaries. Sovan Biswas and R. Venkatesh Babu. Video Analytics Lab, Indian ...

Anomaly Detection via Online Over-Sampling Principal Component ...
... detection has been an important research topic in data mining and machine learning. Many real- ... detection in cyber-security, fault detection, or malignant ..... Anomaly Detection via Online Over-Sampling Principal Component Analysis.pdf.

Multimodal Execution Monitoring for Anomaly Detection ...
Multimodal Execution Monitoring for Anomaly Detection. During Robot Manipulation. Daehyung Park*, Zackory Erickson, Tapomayukh Bhattacharjee, and Charles C. Kemp. Abstract—Online detection of anomalous execution can be valuable for robot manipulati

Anomaly Detection and Attribution in Networks with ...
Abstract—Anomaly detection in communication networks is the first step in the challenging task of securing a net- work, as anomalies may indicate suspicious behaviors, attacks, network malfunctions or failures. In this work, we address the problem

An Anomaly Detection and Isolation Scheme with ...
mean time before a false alarm or a false isolation. Applicability ..... Model-based nuclear power plant monitoring and fault detection: Theoretical foundations.

Anomaly Detection via Optimal Symbolic Observation of ...
Condition monitoring and anomaly detection are essential for the prevention of cascading ..... no false alarm, and no missed detection of special events (such as ...