Speculative Markov Blanket Discovery for Optimal Feature Selection Sandeep Yaramakala Department of Computer Science Iowa State University Ames, IA 50011, USA [email protected] Abstract In this paper we address the problem of learning the Markov blanket of a quantity from data in an efficient manner. Markov blanket discovery can be used in the feature selection problem to find an optimal set of features for classification tasks, and is a frequently-used preprocessing phase in data mining, especially for high-dimensional domains. Our contribution is a novel algorithm for the induction of Markov blankets from data, called Fast-IAMB, that employs a heuristic to quickly recover the Markov blanket. Empirical results show that Fast-IAMB performs in many cases faster and more reliably than existing algorithms without adversely affecting the accuracy of the recovered Markov blankets.

1. Introduction It is often the case that an engineer or researcher is interested in one particular attribute in a set of observations. To analyze and possibly predict the value of this attribute, he or she needs to first ascertain which of the other attributes in the domain affect it. This task is frequently referred to as the feature selection problem. A solution to this problem is often non-trivial, and can be infeasible when the domain is defined over a large number of attributes. A principled solution to the feature selection problem is to determine a subset of attributes that can “shield” (render independent) the attribute of interest from the effect of the remaining attributes in the domain. Koller and Sahami [4] first showed that the Markov blanket of a given target attribute is the theoretically optimal set of attributes to predict its value. Because the Markov blanket of a target attribute T renders it statistically independent from all the remaining attributes (see the Markov blanket definition below), all information that may influence its value is stored in the values of the attributes of its Markov blanket. Any attribute from the feature set outside its Markov blanket can be effectively ignored from the feature set without adversely af-

Dimitris Margaritis Department of Computer Science Iowa State University Ames, IA 50011, USA [email protected] fecting the performance of any classifier that predicts the value of T . Definition 1 (Markov blanket). A Markov blanket B(T) of an attribute T ∈ U is any subset S of attributes for which (T ⊥⊥ U − S − {T } | S) and T 6∈ S.

(1)

A set is called a Markov boundary of T if it is a minimal Markov blanket of T , i.e., none of its proper subsets satisfy Equation (1). In this paper, we identify the Markov blanket of an attribute with its Markov boundary. We use capitals (e.g., X, Y , etc) to indicate attributes and bold letters to indicate sets of attributes e.g., Z. T is the target attribute whose blanket we are seeking, while B(T) is its Markov blanket (a set of attributes). The notation (X ⊥⊥ Y | Z) denotes that X and Y are conditionally independent given Z. Similarly, (X 6⊥⊥ Y | Z) denotes conditional dependence. rX and rZ denote the number of values that attribute X and the variables in set Z (jointly) take, respectively. The number of records in the data set is denoted as N . U is the set of all attributes in the domain. In this paper we assume that the data were generated by a single faithful directed graphical model (a Bayesian network), that models the domain. A Bayesian network is a statistical model that is capable of graphically representing independencies that hold in a domain [6]. The existence of a faithful Bayesian network (see definition of faithfulness below) implies that the Markov blanket of any attribute in the domain is unique, and can be easily “read off” the network structure: The Markov blanket of an attribute is the set of parents, children and spouses (i.e., parents of common children) as encoded by the graph structure of the Bayesian network. For example, the figure at the beginning of this paragraph shows a Bayesian network consisting of five attributes. The Markov blanket of Age

Gender

Exposure to Toxics

Smoking

Cancer

Serum Calcium

Lung Tumor

the attribute Cancer is the set {Exposure to Toxics, Smoking, Serum Calcium, Lung Tumor} (the gray nodes in the figure). This set shields Cancer from the effects of those attributes outside it.

idea is that IAMB and its variants might perform better because (hopefully) fewer false positives will be added during the growing phase (that would have had to be removed during the shrinking phase).

Definition 2 (Faithfulness). A Bayesian network B and a joint distribution P are faithful to to one another iff every conditional independence entailed by the graph of B is also present in P i.e.,

3. The Fast-IAMB Algorithm

(X ⊥⊥B Y | Z) ⇐⇒ (X ⊥⊥P Y | Z) Theorem 1. If a Bayesian network B is faithful, then for every attribute T , B(T) is unique and is the set of parents, children and spouses of T . The goal of this paper is to develop a fast algorithm for discovering Markov blankets from data. We emphasize that we do not address Bayesian network structure discovery here—Markov blankets are discovered without determining the structure of the underlying Bayesian network.

2. Related Work Margaritis and Thrun [5] presented the first provably correct algorithm that discovers the Markov blanket of an attribute from data, under assumptions (see below). The Grow-Shrink Markov Blanket algorithm (GSBN), a Bayesian network structure induction algorithm, invokes the Grow-Shrink Markov blanket algorithm (called GSMB) for every attribute in the domain as a first step. It then utilizes knowledge of the Markov blankets recovered to make the actual Bayesian network structure discovery more efficient. As implied by its name, the GS Markov blanket algorithm contains two phases: a growing phase and a shrinking phase. The GSMB algorithm has the desirable property that, under assumptions, is provably sound i.e., it can recover the exact Markov blanket of any given attribute in the domain. The assumptions made are: (i) the existence of a faithful Bayesian network for the domain under consideration (this implies the existence and uniqueness of the blanket, see Theorem 1 above) and, (ii) the assumption that the conditional independence tests are correct. Tsamardinos, Aliferis, and Statnikov [7] describe a number of variants of GSMB that aim at improved speed and reliability. We evaluate here the Incremental Association Markov Blanket (IAMB) and Interleaved IAMB (Inter-IAMB) algorithms. Like GSMB, the IAMB and Inter-IAMB algorithms also use a two-phase approach for discovering Markov blankets. However, they reorder the set of attributes each time a new attribute enters the blanket in the growing phase. This reordering is done using an information-theoretic heuristic function h (conditional mutual information). The motivating

In this section we present a new algorithm for Markov blanket discovery, called Fast-IAMB. The Fast-IAMB algorithm is shown in Fig. 1. 1: B(T ) ← ∅ 2: S ← {A | A ∈ U − {T } and A 6⊥ ⊥ T} 3: while (S = 6 ∅) do 4: hX1 , . . . , X|S| i ← S sorted according to h 5: insufficient data ← FALSE 6: /* Growing phase. */ 7: for i = 1 to |S| do 8: if r ×r N×r > k then 9: 10: 11: 12: 13: 14: 15: 16: 17: 18: 19: 20: 21: 22: 23: 24: 25: 26:

Xi

T

B(T )

B(T ) ← B(T ) ∪ {Xi } else insufficient data ← TRUE goto 15 /* Insufficient data. */ end if end for /* Shrinking phase. */ for each attribute A ∈ B(T ) do if (A ⊥ ⊥ T | B(T ) − {A}) then B(T ) ← B(T ) − {A} end if end for if insufficient data = TRUE and [no attributes were removed in the shrinking phase] then halt else S ← {A | A ∈ U − {T } − B(T ) and (A 6⊥ ⊥ T | B(T ))} end if end while

Figure 1. The Fast-IAMB algorithm.

Similar to GSMB, IAMB, and Inter-IAMB, Fast-IAMB contains a “growing” phase (in which it attempts to add attributes to the blanket B(T)), followed a “shrinking” phase (in which it attempts to remove as many irrelevant attributes as possible). During the growing phase of each iteration, it sorts the attributes that are candidates for admission to B(T ) from most to least conditionally dependent, according to a heuristic function h. (This is similar to IAMB and Inter-IAMB; however, Fast-IAMB uses the more statistically appropriate significance of a G2 conditional statistical test for h rather than the raw conditional information value, as IAMB and Inter-IAMB do). Each such sorting step is potentially expensive, since it involves the calcula-

tion of the G2 test value between T and each member of S: Each such calculation is equivalent to a conducted conditional independence test. The key idea behind Fast-IAMB is to reduce the number of such tests by adding not one, but a number of attributes at a time after each reordering of the remaining attributes following a modification of the Markov blanket. Fast-IAMB speculatively adds one or more attributes of highest G2 test significance without re-sorting after each modification as IAMB and Inter-IAMB do, which (hopefully) adds more than one true members of the blanket. Thus, the cost of re-sorting the remaining attributes after each Markov blanket modification can be amortized over the addition of multiple attributes. The Fast-IAMB algorithm is sound in that it discovers the exact Markov blanket under the same set of assumptions used by existing algorithms viz. the existence of a faithful (though not necessarily known) Bayesian network to the domain under consideration and the assumption that the conditional independence tests performed by the algorithm are correct. The proof of soundness is the same as that given by [7]. A natural question is to determine how many attributes should be added to the blanket at each iteration. We use the following heuristic: we add dependent attributes as long as the conditional independence tests are reliable i.e., we have enough data for conducting them. For this purpose, we use a numeric parameter k that denotes the minimum average number of instances per cell of a contingency table that should be present for a conditional independence test to be deemed reliable. Let the next attribute that we consider for addition to B(T) be X. To perform a reliable conditional independence test between T and X given B(T), the average number of instances per cell of the contingency table of {X, T } ∪ B(T ) must be at least k, i.e., N > k. (2) rT × rB(T ) × rX In all our experiments we choose k = 5 because, as suggested by Agresti [1], this is the minimum average number of instances per cell for the G2 statistic to have a χ2 distribution, which is a requirement for a significance (p-value) to be calculated. Also, note that in lines 7–14 no conditional independence tests are actually performed. The average number of instances per cell (line 8) calculation can be done in constant time. One practical question remains: What is to be done if the average number of instances per cell for each remaining attributes is less than k? Tsamardinos and Aliferis [7] do not refer to this important practical question in their description of IAMB and Inter-IAMB. One has two choices: assume dependence or assume independence. While assuming dependence might seem to be the “safe” choice, in practice this would result in large blankets that are hard to justify and of little practical use. We therefore assume inde-

pendence when the condition in Eq. (2) fails and halt (line 22) returning the current blanket. This does not adversely impact the performance of Fast-IAMB compared to IAMB and Inter-IAMB, as our experiments confirm.

4. Experimental Results In order to empirically compare the performance of FastIAMB with the other Markov blanket discovery algorithms, we conducted a number of experiments on both synthetic and real-world data sets, listed below. Dataset HAILFINDER20K ADULT CENSUS-INCOME

No. of attributes 56 9 34

No. of records 20,000 45,222 142,521

HAILFINDER20K is a synthetic data set, while both ADULT [3] and CENSUS-INCOME [2] are well-known real-world data sets, containing demographic information. The confidence level of each independence test was set to 95% (α = 0.05). For each experiment, we report the number of conditional independence tests conducted to discover the blanket of each attribute of the domain, the total execution time taken by each algorithm, the distribution of the conditioning set sizes of the tests conducted, and a distance measure that indicates the “fitness” of the discovered blanket. We define the latter to be the average, over all attributes X outside the blanket, of the expected KL-divergence between Pr(T | B(T )) and Pr(T | B(T ) ∪ {X}). We can expect this measure to be close to zero when B(T ) is an approximate blanket. This measure is similar to the one proposed by [4]. Fig. 2 (top row) shows that, in almost all cases, FastIAMB requires fewer conditional independence tests than either IAMB or Inter-IAMB. The number of tests directly influences the execution time of each algorithm (as expected), shown in Fig. 3. From this figure one can verify that Fast-IAMB executes faster than both IAMB and InterIAMB for all data sets: The running time of Fast-IAMB ranges from 68% to 82% of the execution time of IAMB, and 52% to 72% of Inter-IAMB. Fig. 2 (middle row) shows that the blankets discovered by Fast-IAMB are approximately as good as IAMB and Inter-IAMB, as measured by the expected conditional KLdivergence between T and all attributes outside the blanket. This allows the blankets that are discovered Fast-IAMB to be used in comparable situations as IAMB and Inter-IAMB. Fig. 2 (bottom row) shows the distribution of the sizes of the conditioning sets, where size is measured as the number of attributes in the conditioning set. In general, conditioning is undesirable since it typically results in less reliable independence tests. As can be seen from the figure, while the numbers of unconditional tests that all three algorithms are comparable, Fast-IAMB conducts significantly fewer con-

200 150 100 50

0

10

20

30

40

50

60

Variable index

IAMB InterIAMB FastIAMB

30 25 20 15 10

0

1

2

0.05 0.04 0.03 0.02 0.01 0

0

10

20

30

40

50

7

0.04 0.03 0.02 0.01 0

60

0

1

2

3

4

5

6

7

ADULT: Distribution of conditioning set sizes 100

IAMB InterIAMB FastIAMB

80

1500 1000

4

5

40 20

0

6

Conditioning set size (number of attributes)

10

15

20

25

30

35

Variable index

IAMB InterIAMB FastIAMB

0.8 0.6 0.4 0.2

0

5

10

15

20

25

30

35

IAMB InterIAMB FastIAMB

1000

70 60 50 40

10

5

1200

800 600 400 200

20 3

60

Variable index

30

500 2

80

CENSUS-INCOME: Distribution of conditioning set sizes

Frequency

Frequency

2000

1

100

0

8

IAMB InterIAMB FastIAMB

90

2500

0

120

CENSUS-INCOME: Average KL-divergence

HAILFINDER20K: Distribution of conditioning set sizes

0

IAMB InterIAMB FastIAMB

140

1

Variable index

3000

8

IAMB InterIAMB FastIAMB

0.05

Variable index 3500

Frequency

6

CENSUS-INCOME: Number of tests 160

ADULT: Average KL-divergence

Average KL-divergence

Average KL-divergence

0.06

5

0.06

IAMB InterIAMB FastIAMB

0.07

4

Variable index

HAILFINDER20K: Average KL-divergence 0.08

3

Average KL-divergence

250

ADULT: Number of tests 35

Number of conditional independence tests

IAMB InterIAMB FastIAMB

Number of conditional independence tests

Number of conditional independence tests

HAILFINDER20K: Number of tests 300

0

1

2

3

4

Conditioning set size (number of attributes)

0

0

1

2

3

4

5

6

Conditioning set size (number of attributes)

Figure 2. (Top row ): Number of conditional independence tests for each attribute’s blanket discovery in each data set. (Middle row ): Average expected conditional KL-divergence, measuring the fitness of the recovered Markov blankets. (Bottom row ): Distribution of conditioning set sizes of conditional tests conducted.

Total execution time 140 Execution time (sec)

120

IAMB InterIAMB FastIAMB

100 80 60 40 20 0 HAILFINDER20K

ADULT Data Set

CENSUS-INCOME

Algorithm Exec. time (sec) HAILFINDER20K data set IAMB 78.09 Inter-IAMB 82.52 Fast-IAMB 57.83 ADULT data set IAMB 1.71 Inter-IAMB 2.21 Fast-IAMB 1.16 CENSUS-INCOME data set IAMB 109.40 Inter-IAMB 123.17 Fast-IAMB 89.58

Figure 3. Total execution times for each data set. ditional tests compared to IAMB and Inter-IAMB, which indicates improved test reliability.

5. Conclusion and Future Research The main contribution of this paper is a novel algorithm for the induction of Markov blankets from data, called FastIAMB, that employs speculation to recover the Markov blanket faster. Our empirical results show that Fast-IAMB performs often faster and more reliably than existing algorithms, without adversely affecting the accuracy of the recovered Markov blankets. A direction of potential future research is relaxing the requirement of existence of a faithful

underlying Bayesian network (which can be difficult to ascertain in practice) while maintaining the theoretical optimality of the recovered Markov blanket with respect to feature selection.

References [1] A. Agresti. Categorical Data Analysis. New York: John Wiley and Sons Inc., 1990. [2] S. Hettich and S. D. Bay. The UCI KDD archive, 1999. [http://kdd.ics.uci.edu/] UC Irvine, Dept. of ICS. [3] S. Hettich, C. Blake, and C. Merz. UCI repository of machine learning databases, 1998. [http://www.ics.uci.edu/∼mlearn/MLRepository.html] UC Irvine, Dept. of ICS. [4] D. Koller and M. Sahami. Toward optimal feature selection. In International Conference on Machine Learning, pages 284– 292, 1996. [5] D. Margaritis and S. Thrun. Bayesian network induction via local neighborhoods. In S. Solla, T. Leen, and K.-R. Mu¨ ller, editors, Proceedings of Conference on Neural Information Processing Systems (NIPS-12). MIT Press, 1999. [6] J. Pearl. Probabilistic reasoning in intelligent systems: Networks of plausible inference. Morgan Kaufmann Publishers Inc., 1988. [7] I. Tsamardinos, C. Aliferis, and A. Statnikov. Algorithms for large scale Markov blanket discovery. In The 16th International FLAIRS Conference, St. Augustine, Florida, USA, 2003.

Speculative Markov Blanket Discovery for Optimal ...

It is often the case that an engineer or researcher is inter- ested in one particular ... ily “read off” the network structure: The Markov blanket of an attribute is the set ...

202KB Sizes 4 Downloads 241 Views

Recommend Documents

Speculative Markov Blanket Discovery for Optimal Feature Selection
the remaining attributes in the domain. Koller and Sahami. [4] first showed that the Markov blanket of a given target at- tribute is the theoretically optimal set of ...

Local Learning Algorithm for Markov Blanket Discovery
Ecole Polytechnique de Montreal,. C.P.6079, Succ. Centre-ville, Montreal, Quebec, Canada ... show that (i) BFMB significantly outperforms IAMB in measures of data ...... Spirtes, P., Glymour, C.: An algorithm for Fast Recovery of Sparse Casual ...

Fast Markov blanket discovery by Sandeep Yaramakala ...
The discovery of the Markov blanket is useful in the feature subset ...... The Ling-spam corpus contains discussion email messages on a linguistics mailing list ...

Fast Markov Blanket Discovery Algorithm Via Local Learning within ...
is further increased through collecting full statistics within a single data pass. We conclude that IPC-MB plus AD-tree appears a very attractive solution in very large applications. Keywords: Markov blanket, local learning, feature selection, single

Markov Blanket Feature Selection for Support Vector ...
ing Bayesian networks from high-dimensional data sets are the large search ...... Bayesian network structure from massive datasets: The “sparse candidate” ...

A Lifetime Optimal Algorithm for Speculative PRE
network problems. General Terms: Algorithms, Languages, Experimentation, Performance. Additional Key Words and Phrases: Partial redundancy elimination, classic PRE, speculative. PRE, computational optimality, lifetime optimality, data flow analysis.

Optimal risk control and investment for Markov ...
Xin Zhanga †. Ming Zhoub. aSchool of ... Di Masi, Kabanov and Runggaldier (1994), Yao, Zhang, and Zhou (2001), Buffington and. Elliott (2002), Graziano and ...

Convex Synthesis of Optimal Policies for Markov ...
[11], automatic control [12], flight control [13], economics. [14], revenue ... Emails: [email protected] and [email protected] through a sequence of repeated experiments. ...... can send the vehicle to “left” with some probability). For

Blanket Permission Slip.pdf
Students may be transported by staff in. staff vehicles, school issued vans, or school buses. Details of field trips ... permission to. participate in school field trips throughout the 2016-17 school year. Parent Signature: Date: ______. Page 1 of 1.

Application of the Optimal Discovery Procedure to ...
Mar 10, 2011 - Fax +41 61 306 12 34. E-Mail [email protected] ... +44 1223 767 408, E-Mail ioanna.tachmazidou @ mrc-bsu.cam.ac.uk. © 2011 S. Karger AG, ..... lele frequency model with one free parameter, say 1j. We then define 1j { 1j + ...

Speculative Clustered Caches for Clustered Processors
2 Yale University, Department of Computer Science. New Haven, CT, 06520, USA. Abstract. Clustering is a technique for partitioning superscalar pro- cessor's ...

Extend blanket contracts for appraisal services Brian Eades.pdf
Page 2 of 2. VII(d) Extend blanket contracts for appraisal services Brian Eades.pdf. VII(d) Extend blanket contracts for appraisal services Brian Eades.pdf. Open.

Blanket Purchase order.pdf
Whoops! There was a problem loading more pages. Whoops! There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Blanket Purchase order.pdf. Blanket Purchase order.p

7(d) On-Call Blanket Contracts for geo-environmental consulting ...
Page 3 of 3. 7(d) On-Call Blanket Contracts for geo-environmental consulting and construction materials testing $400,000.pdf. 7(d) On-Call Blanket Contracts for ...

Discovery-for-Credit_History.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item.

A Speculative Control Scheme for an Energy-Efficient Banked ... - Scale
energy, and delay numbers from detailed circuit layouts in. Section 3. We present ..... have explored alternative designs for implementing a large ..... Technology.

A Speculative Control Scheme for an Energy-Efficient ...
The trend toward simulta- ... approaches retain a centralized microarchitecture, but divide ..... remove the conflict by setting enable signals such that a local port ...

OPTIMAL RESOURCE PROVISIONING FOR RAPIDLY ...
OPTIMAL RESOURCE PROVISIONING FOR RAPIDL ... UALIZED CLOUD COMPUTING ENVIRONMENTS.pdf. OPTIMAL RESOURCE PROVISIONING FOR ...

Markov Logic
MAP INFERENCE IN MARKOV LOGIC. NETWORKS. ○ We've tried Alchemy (MaxWalkSAT) with poor results. ○ Better results with integer linear programming.