IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 218-223

International Journal of Research in Information Technology (IJRIT) www.ijrit.com

ISSN 2001-5569

Survey on clustering of uncertain data 1

Bessy Basil1, Prof.M.K.Sulaiman2 PG Scholar, Department Of Computer Science and Engineering, TKMCE, University Of Kerala , Kerala, India [email protected]

2

Professor, Department Of Computer Science and Engineering, TKMCE, University Of Kerala , Kerala, India [email protected]

Abstract In the recent years there is an increasing quantity of data with uncertainty being accumulated due to indirect methods of data collection methods like using sensor network, mobile applications, forecasting and imputations. Modeling and mining on such databases with probabilistic information is complex and challenging. This uncertainty is typically modeled using probability. Clustering is a prominent mining operation that stands as a base for applying several other data mining applications. This paper mainly discuses on different models of uncertain data and feasible methods for clustering data with uncertainty.

Keywords: Uncertain data, Data Mining, Clustering.

1. Introduction Data from many natural and social phenomena are accumulated into huge databases in the worldwide network of computers. It is difficult to analyze this vast amount of data by hand since data are massive and complex. Thus the goal is to obtain valuable knowledge by using advanced data analysis mechanisms that exploit the computing power available today. Uncertainty is generally caused by limited perception or understanding of reality, limited observation equipment, limited resources to collect, store, transform, analyze, and understand data. Sensors used to collect data may be thermal, electromagnetic, chemical, mechanical, optical radiation or acoustic used in security, environment surveillance, manufacture systems. Ideal sensors are linear and sensitive, such that the output signal is linearly proportional to the value of the examined property. Practically due to changing environmental conditions ideal sensors outputs cannot be expected. Aggregation of data and granularity of data are also contributing to uncertainty in data. The data we handle have uncertainties in many cases. One of the most general cases of uncertainty is the errors being introduced when the object is mapped from the real space to the pattern space, For example, the rounding error is introduced when we map a data from the real space to the pattern space. When a spring scale of which the measurement accuracy is ±2g shows 100g, an actual value is in the interval from 98g to 102g. Missing values in data also introduces uncertainty. In case of a social investigation, if there are unanswered items in the questionnaire, the items are handled as missing values. Uncertainty may be introduced in real space. In case of measuring a scale about an astronomical object, the estimated scale may have many errors because of the uncertainty such as the fluctuation by measuring gage, the lack of knowledge about the object and so on. Another example of original uncertainty with data Bessy Basil,IJRIT

218

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 218-223

can be analyzed from the field of image. Let’s consider to map some object from the real space to the three dimensional RGB color space. In general, the object in the pattern space is represented as one point. But actually, the object has some colors so that it is more natural that the object is represented as not one point but an interval in the pattern space.

1.1 Modeling of uncertain data Modeling uncertain data is widely addressed problem. Basically uncertainty can be at two perspectives namely existential uncertainty and attribute level uncertainty. Existential uncertainty is the absence of certain data points in a dataset; this may lead to wrong data summarization and incorrect pattern mining. The presence or absence of one tuple may affect the probability of presence or absence of another tuple. Attribute level uncertainty is more difficult to address, it indicates the vagueness in the specification of certain attributes in a data set, as a result data item in such a data will possess corresponding ambiguity. The uncertainties of individual attributes are modeled by a probability density function, or other statistical parameters such as variance. Another view of classification of data uncertainty is in the form of relational uncertainty and spatial uncertainty Probability interpretations of the inherent uncertainties in data may be based on frequentists which is the relative frequency of the occurrences of an outcome, when the experiment is repeated. Another parameter is subjectivness expressed in terms of degree of belief assigned by a person. Interpretations based on Bayesian [15] theory include expert knowledge and experimental data to generate probabilities. Prior probability distribution is used to represent the expert knowledge and likelihood function incorporates the data. The product of the prior probability and the likelihood probability when normalized furnishes a posterior probability distribution which consolidates all of the available information. This paper is organized as follows: In Section 2, we will explore the issue of uncertain data representations and modeling. We will examine data management algorithms for uncertain data particularly in the field of clustering. Section 3 contains the conclusions and summary.

2 Literature review Almost all mining operations on the traditional data bases are equally relevant in uncertain databases. The main uncertain mining operations available in literature are will be clustering, classification, frequent pattern mining and outlier detection. Clustering algorithms on uncertain data can be broadly classified into partitioning clustering algorithms, density based clustering algorithms, clustering based on possible world’s semantics, Computational geometric clustering algorithms and approximation algorithm for clustering.

2.1 Partitioning Clustering Algorithms Most of the partitioning algorithms are based on the K-means clustering algorithm where the main aim is to minimize the sum of squared error. Partitioning algorithms on uncertain data possess a main challenge of expensive distance calculations, Hence the main aim is to reduce the computational complexity by using pruning operations or by using fast approximations. The uncertain version of classic K-Means algorithm, UK-means [10] starts by randomly selecting k points as cluster representatives. Each object Oi is then assigned to the cluster whose representative pj has the smallest expected distance from Oi among all clusters. Next the mean of the centers of the mass of the assigned objects are the new cluster representatives. ED(oi, pj)=∫ fi(x)d(x, pj)dx

(1)

If UK-means iterates t times for a set of n objects to form k clusters, UK-means would compute a total of nkt expected distances. Expected Distance calculation in computationally expensive

Bessy Basil,IJRIT

219

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 218-223

Wang Kay Ngai et al [10] use inexpensive distance calculations to identify cluster representatives that cannot be the closest one to an object. Hence, the expected distances between those representatives and the object need not be computed. More specifically, for each object oi, we define a minimum bounding rectangle (MBR) outside which the object has zero probability of occurrence .Now, for each cluster representative pj, we compute the minimum distance (MinDistij) and maximum distance (MaxDistij) between pj and the MBR of oi . Among all the maximum distances, the smallest one is called the min-max distance di between oi and the cluster representatives. Any remaining cluster representative pj can be pruned if MinDistij > di For others the expected distances from object Oi are calculated. The main problems encountered while following this approach is that we are only considering the centers of objects into account. In the case of uncertain data, as every object has the same center, the expected distance-based approaches cannot distinguish the two sets of objects having different distributions.

2.2 Density Based approaches A cluster is a region of dense points which are surrounded by sparse regions. Density based approaches are used when the clusters are irregular or intertwined and when noise and outliers are present. It tries to identify those dense (highly populated) regions of the multidimensional space and separate them from other dense regions. In the following definitions, a database D with set of points of k- dimensional space S has been used. As we need to find out the object neighbors which are exist/surrounded with in the given radius (Eps), Euclidean function dist (p, q) has been used, where p and q are the two objects. This function takes two objects and gives the distance between them. Definition 1: Eps Neighborhood of an object p - The Eps Neighborhood of an object p is referred as NEps(p), defined as NEps(p) = {q | dist(p,q) <=Eps}

(2)

Definition 2: Core Object Condition - An Object p is referred as core object, if the neighbor objects count >= given threshold value (MinObjs). i.e |NEps(p)|>=MinObjs

(3)

Where MinObjs refers the minimum number of neighbor objects to satisfy the core object condition. In the above case, if p has neighbors which are existing within the Eps radius and if their count is >= MinObjs, p can be referred as core object. Definition 3: Directly Density Reachable Object - An Object p is referred as directly density reachable from another object q w.r.t Eps and MinObjs if p ϵNEps(q) and |NEps(q)|>= MinObjs (Core Object condition)

(4)

Definition 4: Density Reachable Object - An object p is referred as density reachable from another object q w.r.t Eps and MinObjs if there is a chain of objects O1,…,On, O1=q, On=p such that Oi+1 is directly density reachable from pi. Definition 5: Density connected object - An Object p is density connected to another object q if there is an object o such that both, p and q are density reachable from o w.r.t Eps and MinObjs. Definition 6: Cluster - A Cluster C is a non-empty subset of a Database D w.r.t Eps and MinObjs which satisfying the following conditions. 1. For every p and q, if p ϵ cluster C and q is density reachable from p w.r.t Eps and MinObjs then q ϵ C. 2. For every p and q, q C; p is density connected to q w.r.t Eps and MinObjs. Definition 7: Noise - An object which doesn’t belong to any cluster is called noise. Bessy Basil,IJRIT

220

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 218-223

The DBSCAN [6] algorithm first finds the Eps Neighborhood of each object in a database during , then algorithm marks the non-core objects as noise. Every core object is initially treated as a cluster and its neighbors will be the candidates for further cluster expansion. Further the Eps Neighborhood of each candidate object is analyzed. When the new object is a core object, all its neighbor objects will be assigned with the current cluster id and its unprocessed neighbor objects will be pushed into queue for further processing. This cluster expansion process continues until the candidate list is empty. Another related technique discussed is that of hierarchical density based clustering. An effective (deterministic) density based hierarchical clustering algorithm is OPTICS [13]. We note that the core idea in OPTICS is quite similar to DBSCAN and is based on the concept of reachability distance between data points. While the method in DBSCAN defines a global density parameter which is used as a threshold in order to define reachability, the work in points out that different region in the data may have different data density, as a result of which it may not be possible to define the clusters effectively with a single density parameter. Rather, many different values of the density parameter define different (hierarchical) insights about the underlying clusters. The goal is to define an implicit output in terms of ordering data points, so that when the DBSCAN is applied with this ordering, one can obtain the hierarchical clustering at any level for different values of the density parameter. The key is to ensure that the clusters at different levels of the hierarchy are consistent with one another. One observation is that clusters defined over a lower value of ϵ are completely contained in clusters defined over a higher value of ϵ, if the value of MinPts is not varied. Therefore, the data points are ordered based on the value of ϵ required in order to obtain MinPts in the ϵneighborhood. If the data points with smaller values of ϵ are processed first, then it is assured that higher density regions are always processed before lower density regions. This ensures that if the DBSCAN algorithm is used for different values of with this ordering, then a consistent result is obtained. Thus, the output of the OPTICS algorithm is not the cluster membership, but it is the order in which the data points are processed. We note that that since the OPTICS algorithm shares so many characteristics with the DBSCAN algorithm, it is fairly easy to extend the OPTICS algorithm to the uncertain case using the same approach as that was used for extending the DBSCAN algorithm. This is referred to as the FOPTICS algorithm. Note that one of the core-concepts needed to order to data points is to determine the value of ϵ which is needed in order to obtain MinPts in the corresponding neighborhood. In the uncertain case, this value is defined probabilistically, and the corresponding expected values are used to order the data points. For analyzing uncertain data using standard OPTICS algorithm the distance between data objects need to be represented in one numeral value. The single valued numeral cannot model clearly the uncertainty in data. The fuzzy version of the DBSCAN algorithm (referred to as FDBSCAN) works in a similar way to the DBSCAN algorithm, except that the uncertainty of the database makes the corresponding density at each point also uncertain. This corresponds to the fact that the number of data points within the ϵ-neighborhood of a given data point can be estimated only probabilistically, and is essentially an uncertain variable. Correspondingly, the reachability from one point to another is no longer deterministic, since other data points may lie within the ϵ-neighborhood of a given point with a certain probability. Hans-Peter Kriegel et al[13] models the similarity between fuzzy objects by using distance probability functions. It assigns a probability value to each of the possible distance value. Unlike to traditional approach, [13] do not try to extract aggregated values from fuzzy distance functions but enhances OPTICS to utilize the full information revealed by these functions. Thus the user gets an overview over a large set of fuzzy objects using this algorithm.

2.3 Possible world approaches In the possible world’s semantics [5], a set of possible worlds are sampled from an uncertain data set. Each possible world consists of an instance from each object. Clustering is conducted individually on each possible world .Clusters of each possible world is aggregated to obtain the final clusters. The sum of the difference between the global clustering and the clustering of every possible world should be the minimum. Bessy Basil,IJRIT

221

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 218-223

The main problem with possible worlds approach for uncertain data is that each possible world only contains one instance from each object and thus it cannot clearly depict the true distribution of a data object. The clustering results from different possible worlds can be drastically different also the most probable clusters calculated using possible worlds may still have a very low probability. The additional challenge is that it is computationally infeasible due to the exponential number of possible worlds.

2.4 Computational geometric clustering algorithms The work in [10] uses minimum bounding boxes of the uncertain objects in order to compute the distance bounds for effective pruning. It makes use of the concept of voronoi diagram where each cell in the voronoi diagram is associated with a cluster representative. If the minimum bounding rectangle of an uncertain object lies completely inside a cell, then it is not necessary to compute its distance to any other cluster representatives. For any pair of cluster representatives, the perpendicular bisector between the two is a hyper plane which is equidistant from the two representatives by the property of voronoi diagram. In the event that the MBR of an uncertain object lies completely on one side of the bisector, we can deduce that one of the cluster representatives is closer to the uncertain object than the other. This allows us to prune of the representatives.

2.5 Approximation algorithms for uncertain clustering Recently, techniques have been designed for approximation algorithms for uncertain clustering. The work in [14] discusses extensions of the k-mean and k-median version of the problems. Bi-criteria algorithms are designed for each of these cases. A key approach proposed in the paper [14] is the use of a transformation from the uncertain case to a weighted version of the deterministic case. We note that solutions to the weighted version of the deterministic clustering problem are well known, and require only a polynomial blow-up in the problem size. The key assumption in solving the weighted deterministic case is that the ratio of the largest to smallest weights is polynomial. This assumption is assumed to be maintained in the transformation. This approach can be used in order to solve both the uncertain k-means and k-median version of the problem.

3. Conclusions The advances in data collection and data storage have led to the need for processing data with uncertainty. Almost all of the classical data mining algorithms need to be reformulated in order to adapt to the inherent uncertainty in data. This paper provides a survey on modeling and clustering techniques on uncertain data and identifies high dimensionality of data as an additional challenge to uncertainty. Possible future research work includes studies against clustering data streams, statistics estimates, spatial, temporal and graphical data with uncertainty.

References [1] Bin Jiang, Jian Pei, Yufei Tao and Xuemin Lin (2013), “Clustering Uncertain Data Based on Probability Distribution Similarity”, IEEE transactions on knowledge and data engineering, 25(4), 751763 . [2] Dan Olteanu(2012), “Clustering Correlated Uncertain Data”, KDD’12, ACM 978-1-4503-1462-6 [3] Kriegel, H.-P., & Pfeifle, M. (2005), “Hierarchical density-based clustering of uncertain data”, Fifth IEEE International Conference on Data Mining (ICDM’05).

Bessy Basil,IJRIT

222

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 218-223

[4] Song, Q., Ni, J., & Wang, G. (2013),” A Fast Clustering-Based Feature Subset Selection Algorithm for High-Dimensional Data”, IEEE Transactions on Knowledge and Data Engineering, 25(1), 1–14. [5] S. Abiteboul, P.C. Kanellakis and G. Grahne,(1987), “On the Representation and Querying of Sets of Possible Worlds,” Proc. ACM SIGMOD Int’l Conf. Management of Data (SIGMOD) [6] Ester M., Kriegel H.-P., Sander J., and Xu X. (1996) , “A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise”, 2nd International Conference on Knowledge Discovery and Data Mining (KDD‟96), 226-231 [7] Olteanu, D. (2012),” DAGger : Clustering Correlated Uncertain Data (to predict asset failure in energy networks )”, KDD’12, ACM, 0–3. [8] Mostafaei, H. (2011). Probability Metrics and their Applications, Applied Mathematical Sciences ,5(4), 181–192. [9] Stephan, G. Kremer, H. Hardy Kremer ,Thomas Seidl,. “Subspace Clustering for Uncertain Data”, SIAM ,385–396. [10] Ngai, W. K. N. W. K., Kao, B. K. Ben, Chui, C. K. C. C. K., Cheng, R., Chau, M., & Yip, K. Y. (2006),” Efficient Clustering of Uncertain Data”, Sixth International Conference on Data Mining (ICDM’06). [11] Aggarwal, C. C., Member, S., & Yu, P. S. (2009),” A Survey of Uncertain Data Algorithms and Applications”, IEEE transactions on knowledge and data engineering, 21(5), 1–15. [12] Dixit, A., & Misal. A. (2012),”A Survey on Uncertain Data & its Clustering”, International Journal of Computer Science and Management Research, 1(4),736–741. [13] Kriegel, H.-P., & Pfeifle, M. (2005).” Hierarchical density-based clustering of uncertain data”, Fifth IEEE International Conference on Data Mining (ICDM’05). [14] G. Cormode, and A. McGregor, “Approximation algorithms for clustering uncertain data,” PODS Conference, 191-200, 2008. [15] Qin, B., Xia, Y., Wang, S., & Du, X. (2011), “A novel Bayesian classification for uncertain data”, Knowledge-Based Systems, 24(8), 1151–1158.

Bessy Basil,IJRIT

223

Survey on clustering of uncertain data urvey on ...

This paper mainly discuses on different models of uncertain data and feasible methods for .... More specifically, for each object oi, we define a minimum bounding .... The advances in data collection and data storage have led to the need for ...

70KB Sizes 0 Downloads 232 Views

Recommend Documents

Survey on Data Clustering - IJRIT
common technique for statistical data analysis used in many fields, including machine ... The clustering process may result in different partitioning of a data set, ...

Survey on Data Clustering - IJRIT
Data clustering aims to organize a collection of data items into clusters, such that ... common technique for statistical data analysis used in many fields, including ...

A Survey on Data Stream Clustering Algorithms
The storage, querying, processing and mining of such data sets are highly .... problems, a novel approach to manipulate the heterogeneous data stream ...

A survey on enhanced subspace clustering
Feb 6, 2012 - spective, and important applications of traditional clustering are also given. ... (2006) presents the clustering algorithms from a data mining ...

A survey on enhanced subspace clustering
clustering accounts the full data space to partition objects based on their similarity. ... isfy the definition of the cluster, significant subspace clusters that are ...

Protecting sensitive knowledge based on clustering method in data ...
Protecting sensitive knowledge based on clustering method in data mining.pdf. Protecting sensitive knowledge based on clustering method in data mining.pdf.

Survey on Malware Detection Methods.pdf
need the support of any file. It might delete ... Adware or advertising-supported software automatically plays, displays, or .... Strong static analysis based on API.

A Survey on Brain Tumour Detection Using Data Mining Algorithm
Abstract — MRI image segmentation is one of the fundamental issues of digital image, in this paper, we shall discuss various techniques for brain tumor detection and shall elaborate and compare all of them. There will be some mathematical morpholog

A Short Survey on P2P Data Indexing - Semantic Scholar
Department of Computer Science and Engineering. Fudan University .... mines the bound of hops of a lookup operation, and the degree which determines the ...

Survey on Physical and Data Safety for Cellular ...
Here we are using the GPS based data storage by using cloud services. ... products are nothing like the analog phones from many years ago. People could drop ...

Inference on Inequality from Complex Survey Data 9
data from the complexly designed Indian National Sample Survey. Next, we .... these distributions to complex sample design, we make our procedures ...... Ideally, one should consult the sample design document to see what variables are the.

A Short Survey on P2P Data Indexing - Semantic Scholar
Department of Computer Science and Engineering. Fudan University ... existing schemes fall into two categories: the over-DHT index- ing paradigm, which as a ...

Inference on Inequality from Complex Survey Data
testing economic inequality when the data come from stratified and clustered ..... and its estimate are cadlag (continuous from the right with limit on the left), ...

On relational possibilistic clustering
Keywords: Cluster analysis; Possibilistic c-means; Relational data; Dissimilarity measures. 1. ..... tion yields null values for large distances only asymptoti- cally.

Feedback on EBF survey on Incurred Sample Stability (ISS)
Nov 17, 2011 - Why consider ISS analysis? http://www.europeanbioanalysisforum.eu. 5. Scientific. ➢ Back conversion of unknown conjugates – directly on parent or metabolites. ➢ Acyl Glucuronides. ➢ labile analytes / metabolites. ➢ Enzymatic

Impact of volatility clustering on equity indexed annuities
Sep 24, 2016 - On the other hand, empirical analysis conducted in Aït-Sahalia et al. ...... diffusion model (JD) is adjusted to the time series data of S&P 500 ...

A Comparison of Scalability on Subspace Clustering ...
IJRIT International Journal of Research in Information Technology, Volume 2, Issue 3, March ... 2Associate Professor, PSNA College of Engineering & Technology, Dindigul, Tamilnadu, India, .... choosing the best non-redundant clustering.

The Effect of Membrane Receptor Clustering on Spatio ... - Springer Link
clustering on ligand binding kinetics using a computational individual- based model. The model .... If the receptor is free – not already bound to a ligand ...

The Deterrent Effect of Cable System Clustering on ...
Mar 11, 2015 - Headends are the control centers of a cable system. ... “Moreover, we note that data submitted in the record by cable operators indicate that ...

Update on Abyei - Small Arms Survey Sudan
Jan 30, 2015 - In the 20 June 2011 Addis Ababa agreement, the Government of Sudan (GoS) .... open support for the SPLM/A-IO would likely result in swift ...

Update on Abyei - Small Arms Survey Sudan
Jan 30, 2015 - In the 20 June 2011 Addis Ababa agreement, the Government of Sudan (GoS) .... open support for the SPLM/A-IO would likely result in swift ...

Mini survey on settlement hierarchy.pdf
Mini survey on settlement hierarchy.pdf. Mini survey on settlement hierarchy.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying Mini survey on ...

data clustering
Clustering is one of the most important techniques in data mining. ..... of data and more complex data, such as multimedia data, semi-structured/unstructured.