Submitted: IEEE Trans. Patt. Anal. Mach. Intell.

A Multiscale Mean Shift Algorithm for Mode Estimation Lewis D Griffin1 & Martin Lillholm2 1

Computer Science, University College London, UK; [email protected] 2 IT University, Copenhagen, Denmark; [email protected]

We derive and evaluate a novel algorithm for estimating the mode of a distribution based on samples from it. The algorithm is related to the popular mean shift algorithm, but incorporates a principled, intuitive and simple method of changing the scale, as well as the location, of the kernel that is used to estimate the distribution. The strengths of the algorithm are that (i) it is simple, (ii) it is fast, (iii) it lacks data-dependent parameters that need to be set by hand, (iv) it is competitive in RMS estimation error with the state-of-theart for 1-D distributions, (v) it has lower bias, and (v) its multi-dimensional version is as simple and fast as the one-dimensional version, allowing problems to be tackled that cannot be through any previous algorithm

1. Introduction It is often useful to characterize a distribution by identifying a single representative value of the domain. Commonly one uses the mean, which is simple to calculate exactly when given an analytically-specified distribution, and simple to estimate when given only samples from the distribution. However, sometimes the mode is more appropriate, “For example, a medical researcher would normally be more interested in the most probable heart rate than in the total of the heart rate measurements divided by the number of measurements” [1], and unfortunately the mode is simple to calculate but, for continuous data, difficult to estimate. Mode estimation methods have been applied to data from a wide variety of sources and of a wide range of dimensionalities. Starting with the lowest dimensional, 1-D applications include archaeological data [2], measurements of cellular DNA content [3], and time estimates in molecular clock studies [4]. 2-D datasets include seismic tremor locations [5], and 3-D include color histograms [6]. The three dimensions of color have been combined with two dimensions of position, to yield 5-D data which has been analyzed to effect color image segmentation [7]. Jumping to higher dimensional analyses, anatomical growth curves expressed as 20-D vectors have had their modes determined [8]; as have, 48-D image-filter-response vectors [9], 66-D gene expression time-courses [10], and 365-D daily rainfall rates [11]. In this paper we present a novel algorithm for estimating the mode of a distribution given samples from it. The strengths of our algorithm are that (i) it is simple, (ii) it is fast, (iii) it lacks data-dependent parameters that need to be set by hand, (iv) it is competitive in RMS estimation error with the state-of-the-art for 1-D distributions, (v) it has lower bias, and (v) its multidimensional version is as simple and fast as the one-dimensional version, allowing problems to be tackled that cannot be through any previous algorithm. In the remainder of the introduction

we describe previous solutions to the mode estimation problem, paying particular attention to those algorithms related to ours.

1.1 Parametric Methods Mode estimation is straightforward if one has a prior expectation of the exact form of the density (e.g. that it is log-normally distributed) — one simply finds the maximum likelihood estimate of the parameters of the density and then reads off the mode [12]. However, one’s prior expectation is often of only the general rather than the exact form of the distribution. In such a case one needs to find the best fitting member of some suitably expressive family of models. One such approach is the Gaussian mixture method [13] that models a distribution as a sum of Gaussians of varying center and width. The mixture model is fitted to make the observed data as likely as possible and then the mode of the model can be calculated by hill climbing. The mixture model approach can be effective if one knows that the distribution is multi-modal, and one knows the number of modes, but the approach is not effective for uni-modal distributions as these, if skewed, are not compactly modeled as sums of Gaussians. Suitable approaches in these cases are the standard and robust parametric mode (SPM and RPM) estimators [14]. These estimators are based on a family of density models obtained by warping the domain of a normal distribution using a power law transform.

1.2 Kernel-based Methods Lacking an expectation of the form of the density, a way to proceed is to bin the data, plot a histogram and read off its mode. A problem with this is the shift-dependency of the bin counts. Consideration of this problem led to the development of methods where the density is estimated, not by binning, but by convolving the data (represented as a collection of delta functions) with a smoothing kernel [15, 16]. These kernel-based density-estimation methods avoid the shift-dependency of histogram approaches, but by dispensing with distinct bins make locating the mode of the estimated density non-trivial. The solution is to hill-climb the density-estimate. If the kernels used in the estimation process are Gaussians (as is now standard [17]), this is similar to the final step required when using Gaussian mixture models (as described in the previous section), the difference being that rather than there being only a few summed Gaussians, there is one for each sample. Happily, sums of Gaussians can be hill-climbed robustly and efficiently even if there are many in the sum. The hill-climbing method, known as the mean shift [18], is a simple iterative procedure. At each iteration, one considers a Gaussian kernel centered on the current estimate of the mode. This kernel associates a weight with each sample (more distant samples having lower weight). The mean position of the weighted samples is then computed, and the mode estimate is shifted to this position. Shifting is repeated until the movement is small compared to the kernel scale. This mean shift procedure successfully climbs the kernel-estimated density because it is precisely equivalent to a gradient ascent process [7, 19]; moreover the mean shifts are not just in the gradient direction, but actually are a fixed multiple of the gradient of the log of the kernelestimated density [19].

The mean shift procedure is a very important contribution to the theory of mode estimation, but there are two details that need to be supplied for a complete theory. First, is a rule for choosing where to start the mean shift. Admittedly, this issue can be dodged by performing mean shifts starting at different locations (e.g. starting at each different sample), and selecting the final location where the estimated density is maximal, but this is hardly efficient or elegant. The second missing detail is a procedure for deciding what the scale (or synonymously bandwidth) of the Gaussian kernels used should be. Simply trying each of a range of scales does not succeed as a strategy as it is not obvious how to compare the modes found at different scales in order to select the best (fig. 1). The vexed question of what kernel scale is optimum was tackled first by Parzen [16], who proposed to minimize the expected L2 difference between the estimated and true densities, and showed that this depends on the L2-norm of the true density’s 2nd derivative; but since we don’t know the true density this result is not easily used in practice. Since then, numerous other criteria [20] have been proposed, such as (i) assuming a standard form for the true density (e.g. Gaussian [21]), (ii) ‘plug-in methods’ that estimate the density 2nd derivative by assuming normality of its 4th derivative (and so on), (iii) iterative plug-in [22]. Other authors have pursued the idea that a spatially-varying kernel scale is advantageous [6, 23, 24].

Figure 1 – Shows histograms (top-row) and kernel-estimated densities (bottom row) of the same set of samples from a log-normal distribution. Bin-width and kernel-scale decrease from left-to-right. The kernel-scales are chosen to match the bin-widths. Note, that for both kinds of plot the vertical axes have been scaled so that the height of the highest mode is constant. In unscaled plots these heights consistently increase from left-to-right.

An alternative approach to optimum scale selection, more specifically aimed at moderather than density-estimation, is to use bootstrap methods [25]. The idea here is to assess at each candidate scale the precision of the mode estimate by looking at how it varies when using bootstrap resamplings of the data. Informally, very coarse scales will give greatly varying mode estimates because the mode has a broad top, and very fine scales will give greatly varying estimates because the mode depends only on a few samples; somewhere between the two extremes will be a data-dependent optimum scale. The approaches that have been presented [2628] vary in their details, so we will describe fully only the most recent scheme [8]. First a candidate set of scales is chosen. For each candidate scale a mean shift procedure is performed starting at the mean of the data, using kernels of the candidate scale. Then, multiple

times, the data is re-sampled using a smoothed bootstrap [29] (i.e. data-points are drawn randomly-with-replacement from the original set, and are then perturbed by addition of a Gaussian random variable of the scale being assessed) and a mean shift is performed on the resampled set. The RMS error between the raw-data mode-estimate, and the resampled-data modeestimates is calculated. The candidate scale for which this RMS error is minimized is said to be the optimum scale, and the raw-data mode-estimate for that scale is the final overall modeestimate.

1.3 Scale-reduction Methods The key idea behind scale-reduction methods is that mode-estimates at coarse scales can be used to choose between the multiple mode-estimates obtainable at finer scales. For example, for the data in figure 1 (top row) one can imagine a scheme that tracks the modal histogram bin across reducing bin width, so that ‘false’ local modes are ignored and eventually a single sample can be identified as the best mode estimate. This informal idea was first formalized by Bickel [1] in two simple but effective algorithms. The first, the Half Sample Mode, starts by finding the densest half-subset i.e. the subset of the samples that (i) contains half the full number and (ii) covers as short a range as possible. The same process is then applied to this densest half-subset, and so on. Eventually the set being considered has only one or two members, and the mean of this set is returned as the mode estimate. The second algorithm, the Half Range Mode, starts by calculating the range of the data. It then searches for the interval, half the size of the full range, which contains the most data points. The same process is then applied to the points in this densest half-range, and so on. Again, eventually the set being considered has only one or two members, and the mean of this set is returned as the mode estimate. We also previously presented a scale-reduction method of mode estimation [30]. This method makes use of the continuum of density estimates obtained as the kernel scale parameter is varied. Consideration of this continuum has previously been advocated by several authors [3, 31-33] in the density estimation literature, and independently in the Scale Space computer vision literature [34-40]. Common to this prior work is the observation that modes vary continuously in position and annihilate with anti-modes as scale is increased from fine-to-coarse. The proof that using a Gaussian as the kernel has the consequence that, in 1-D, modes are never created as scale increases has been given in both the density estimation [33] and scale space literature [41]. In [30] it is shown that the continuum of density estimates can be used not just as a way to understand and visualize multimodality, but also as a numerical method of mode estimation. The key idea is that, in 1-D, at sufficiently coarse scale there is exactly one mode, and since modes are never created with increasing scale, this mode may be tracked through decreasing scale to one of the original sample points; and this sample point may be taken as an estimate of the mode of the true distribution (we will discuss some caveats to this idea in section 4.1). The emphasis of [30] is a refinement of this basic idea aimed at stopping the scale space tracking at some nonzero scale in a data-dependent way (inspired by the optimum bandwidth literature), while the scale space tracking aspect is rather crude and uses an ad hoc method of balancing domain- vs. scale-shifts (see fig. 2).

Figure 2 – Shows the scale space tracking method of [30]. The samples on which the density estimates are based can be seen along the bottom of the plots. Different cross-sections correspond to different density estimates using kernels of varying scale, from coarse at the top to fine at the bottom. The figure at left shows tracking across regular density estimates, at right across a ‘pessimistic scale space’ designed to prevent tracking to very fine scales. The poor control of the tracking is apparent in the jagged paths.

1.4 Which of the existing methods is best? Since only a few comparative studies of the methods that we consider state-of-the-art have been performed, and since the results tends to vary with the test distributions used, making definitive statements about superiority of different estimators is difficult. In the paper [1] that originally presents the Half Sample Mode (HSM) and Half Range Mode (HRM), the two estimators are compared. In this study the pattern of performance across different test distributions is very similar, but it is concluded that the HRM is slightly more robust. The same author in a later publication [14] compares these two methods to the Standard Parametric Mode (SPM) and Robust Parametric Mode (RPM). The pattern of performance varies, but the author concludes that on balance the RPM is to be preferred. In contrast, a substantial comparison [4] of the HRM, SPM, RPM concludes by advocating the HRM. They note that there is little consistent performance difference between the HRM, SPM & RPM, but that the HRM is vastly simpler and faster. In terms of speed, the HRM and HSM are clearly the best. Next are the SPM and RPM which are much slower as they require a parameter (the coefficient of the power law transform) to be exhaustively explored. Even slower than these methods are bootstrap methods [8] which require a full parameter (scale) to be explored, and multiple bootstraps to be performed at each candidate scale. Also the mean shift processes used in the bootstrap approach can occasionally be very slow when the starting position is very poor (as the mean can be at fine scales).

2. The Multiscale Mean Shift Algorithm In this section we describe, first informally and then rigorously, the multiscale mean shift (MSMS) algorithm. We conclude by giving some detailed notes on implementing the algorithm.

2.1 Informal Description The MSMS algorithm is of the scale-reduction type described in section 1.3. When operating in 1-D it is similar to the HSM and HRM algorithms, but it has the advantage that it generalizes naturally to higher dimensions, which they do not. It is also similar to the Scale Space tracking technique described in the introduction but the ‘pessimistic’ concept that discouraged tracking to very fine scales has been dropped and the correct way to balance domain- and scale-shifts has been discovered. In scale-reduction techniques we aim to follow the path of the mode from coarse- to fine-scale as it varies continuously in position. To do so we can make use of the defining property of the mode path – at each point the density is a within-scale maximum, but this property alone is insufficient for easy mode tracking. In the case that we are using Gaussian kernels for the density estimation, there is an additional property that can be made use of. For these kernels, and no other, the mode path has the property that the density along it is always increasing as one moves to finer scales. Putting these two properties together, if we start near the mode path at its coarse-scale end, then we can approximately track along it by hill-climbing, across domainµscale, the estimated density. Even once a hill climbing approach to mode tracking has been chosen, there are still crucial questions to be answered about how this process should be carried out. The answers to these questions are supplied by considering why the mean shift algorithm is successful, which we now do. It is not immediately obvious what the mean shift algorithm actually contributes to the problem of mode estimation based on density estimation. After all, once the idea of kernel-based density-estimation is in place, using a gradient ascent method to find the mode is obvious. However, it is important to realize that being able to compute the gradient of a function does not automatically make it easy to climb to the maximum. Even in 1-D, gradient ascent requires careful book-keeping to control the step size, so that convergence is achieved efficiently ([42] section 10.3); and in multiple dimensions, not only is even more book-keeping required, but it is even recommended that steps are not always taken in the gradient direction as this can (when combined with imperfect step size selection) lead to zigzagging up ridges ([42] section 10.6). What the mean shift algorithm supplies is a method of calculating the correct multiple of the gradient vector so that a simple scheme of taking such steps robustly climbs to the maximum without book-keeping, zigzagging, oscillation or overshoot. This nice step-size property of the mean shift algorithm is what we require for our MSMS algorithm. The clue to how to achieve this again comes from considering the mean shift algorithm, and noting that the multiple of the gradient vector of the log density that the mean shift algorithm chooses for a shift is also definable in the alternate way as shifting to the mean of the samples as seen under the current kernel. Thus, what we require for MSMS is a scheme that produces a domainµscale-shift vector that is (i) interpretable as some simple operation on the

samples visible under the current kernel, and (ii) is also a multiple of the domainµscale gradient vector of the log estimated-density.

Figure 3 – Illustrates the Multiscale Mean Shift (MSMS) algorithm. At the left is shown the scale space of the set of samples marked with black dots along the bottom of the frame. Each row of this density plot is the kernel-estimated density at a different scale. Scale (i.e. Gaussian kernel width) increases linearly from bottom to top. The true mode path is marked in green. The sequence of positions of the MSMS algorithm are marked with orange and blue dots joined by a red line. Four of the MSMS positions have been chosen for illustrative purposes and have been marked with blue dots. Around these locations are shown frames that extend [-1, 1]2 in the local coordinate frame for that position. These neighborhoods are shown in the four windows at the right, plotted in the local coordinate frames. The density as displayed in the right hand windows has been log-transformed.

The scheme that we formally derive in the following section is that at each iteration of the MSMS the domain shift is to the mean of the samples as weighted by the current kernel, and the scale shift is to the average of the current scale and the scale (i.e. scatter) of the samples as weighted by the current kernel. The method of the derivation is firstly to consider the logtransformed estimated density, and secondly to use a tailored coordinate system at each shift; tailored so that the gradient vector (of the logged density) in that coordinate system is interpretable as operations on the points as visible under the kernel. This is illustrated in figure 3, in which it can be seen that the coordinate systems are such that the local domainµscale neighborhood looks very similar at all points along the mode path. The MSMS algorithm operates unchanged on higher-dimensional data. For example, figure 4 shows it operating on a set of 2D samples.

Figure 4 – Illustrated the MSMS algorithm operating on samples from a 2D distribution. The distribution is shown as a density plot at the left. It is an artificial example, constructed to be clearly non-separable. The right hand panel shows the sequence of mode estimates (black and white curve) and kernel sizes (grey circles) of the MSMS algorithm when applied to the set of 1000 random samples shown in the central panel. The grey dot in the left panel shows the true mode of the distribution. The black dot and ellipse show the mean and 95% quantile of the MSMS mode estimates when applied to sample sets of size 1000.

2.2 Mathematical Derivation The MSMS algorithm tracks through the family of kernel-estimated densities produced using different scales of kernel. The tracking is by a sequence of shifts of location and scale. As with the mean shift algorithm, the shift of location is to the centre of gravity of the samples weighted by the current kernel. The shift of scale is also simply defined in terms of the samples as currently weighted. The first part of our derivation is concerned with showing that given a certain choice of local coordinate scheme for locationµscale, the gradient vector of the log of the estimated density is exactly equal to the shifts defined in terms of the samples as weighted by the current kernel. In the choice of local co-ordinate system, we introduce a single unspecified constant. This constant controls how fast the process tracks to finer scales, just as the ‘half’ aspects of the HSM and HRM algorithms do. The second part of our derivation considers what constraints there are on how fast this scale reduction can and should proceed. The section concludes with our recommendation on the value of this constant.

2.2.1 Derivation of the shift vector We will derive the MSMS algorithm for the general case of samples from an N-dimensional distribution. We will use vector variables (e.g. x , y , χ ) and constants (e.g. pi ) to refer to locations in the domain of the distribution (i.e. N ). We will use positive real-valued variables (e.g. s, ω ) to refer to the scales of kernels, and samples weighted by kernels.

We

define

A ( y; x , s ) := ( 2π s )

− N2

the

e



y−x 2s

Gaussian

kernel

centered

at

location-scale

x, s

as:

2

. Note that, to make the equations of this derivation as simple as

possible, we use a scale variable with dimension LENGTH2, rather than LENGTH as is sometimes used. This means that the scatter (mean distance-from-the-centre-squared) of a kernel is N × s , and so the scale of a kernel is equal to its scatter divided by its dimensionality. The derivative of the kernel with respect to the position of its centre is ( 0,1,0 ) A and with respect to its scale is: ( y ; x , s ) = s −1 ( y − x ) A ( y ; x , s )

(

)

A( 0,0,1) ( y; x , s ) = 12 s −2 y − x − N s A ( y; x , s ) . 2

Let the n sample points from the N-dimensional distribution be denoted p1 ,… , pn ∈

N

. Let the

density as estimated by a kernel be: D ( x , s ) := ∑ A ( pi ; x , s ) . Denote the log density by i



D := ln D .

The centre of gravity of the points observed by a kernel is: 1 C ( x , s ) := ∑ pi A ( pi ; x , s ) D ( x, s ) i We can stretch the concept of scale and apply it to samples as well as kernels. Then, using the scale-equals-scatter-divided-by-dimensionality definition noted above, the scale of the 1 2 samples weighted by a kernel is: ω ( x , s ) := N −1 pi − x A ( pi ; x , s ) ∑ D ( x, s ) i Having set up these definitions, we now derive the location-scale shift from one mode estimate to the next. Let the current mode-estimate be the location-scale χ ,ψ . Define the local coordinate

system

h ( x ) , m ( s ) := ψ

− 12

h , m as

a

function

( x − χ ) , ( 2kN −1 ) (ψ −1s − 1) − 12

of

the

. Where k ∈

fixed +

one

x, s :

is a so-far-unspecified

constant. In this local coordinate system, the coordinates of the current kernel are 0, 0 . The mapping back from the local to the fixed coordinate system is given by:

( )

(

)

x h , s ( m ) = χ + ψ 2 h , 1 + ( 2kN −1 ) 2 m ψ 1

1

and the derivatives of this inverse mapping are:

1 1 ds dx = ψ 2 I N and = ( 2kN −1 ) 2 ψ where I N is the N-dimensional identity matrix. dm dh We now calculate the derivatives of the log density. Note that these derivatives are with respect to the local coordinate system, and are evaluated at the origin of that system (i.e. at the location-scale of the current mode estimate).

(() )

d ∗ D x h ,ψ dh

= h =0

dx d ∗ D ( x ,ψ ) dh dx x=χ 1

d D ( x ,ψ ) D ( χ ,ψ ) dx x=χ

=ψ 2 1

1

=ψ 2 1

=ψ =ψ

D ( χ ,ψ )

− 12

− 12

1

∑ A(

0,1,0 )

( pi ; χ ,ψ )

i

D ( χ ,ψ )

∑ ( p − χ ) A ( p ; χ ,ψ ) i

i

i

(C ( χ ,ψ ) − χ )

d ∗ ds d ∗ = D ( χ , s (m )) D ( χ, s) dm dm ds m =0 s =ψ =( 2kN −1 ) 2 ψ 1

= ( 2kN −1 ) 2 ψ 1

1

d D ( χ, s) D ( χ ,ψ ) ds s =ψ 1

A( ∑ D ( χ ,ψ )

0,0,1)

( pi ; χ ,ψ )

i

= ( kN −1 2 ) 2 ψ −1 1

∑( p − χ D ( χ ,ψ ) 1

i

i

2

)

− Nψ A ( pi ; χ ,ψ )

= ( Nk 2 ) 2 (ψ −1ω ( χ ,ψ ) − 1) 1

The MSMS algorithm is that the shift in location-scale is given by the gradient vector of the log density when expressed in the local coordinate system. We first express this in the local −1 ⎞ C ( χ ,ψ ) − χ ψ 2 ⎛ h′ ⎞ ⎛ 0 ⎞ ⎛ ∂∂h ⎞ * ⎛⎜ ⎟ and then transform it coordinate system: ⎜ ⎟ = ⎜ ⎟ + ⎜ ∂ ⎟ D = 1 ⎜ ⎟ 1 − 2 ⎝ m′ ⎠ ⎝ 0 ⎠ ⎝ ∂m ⎠ ⎝ ( Nk 2 ) (ω ( χ ,ψ )ψ − 1) ⎠ ⎞ C ( χ ,ψ ) ⎛ χ ′ ⎞ ⎛ x h′ ⎞ ⎛ ⎟=⎜ into the fixed system: ⎜ ⎟ = ⎜ ⎟. ⎝ψ ′ ⎠ ⎜⎝ s ( m′) ⎟⎠ ⎜⎝ (1 − k )ψ + k ω ( χ ,ψ ) ⎟⎠ So the location of the new mode estimate is the centre of gravity of the samples as weighted by the old kernel, and the new scale is a weighted sum of the old scale and the scale of the samples as weighted by the old kernel; where the constant k controls the weighting.

(

)

( )

2.2.1 Choosing the rate of scale reduction In the above we showed that the scale shifts to ψ ′ = (1 − k )ψ + k ω ( χ ,ψ ) . So the constant k controls how fast the tracking proceeds to finer scales: smaller k causing slower scale reduction. To arrive at a rule for choosing k we make the following analyses:

Constraint 1 – tracking should be precise The slower the scale reduction, the more accurately the shifting process tracks along the mode path: in the limit k → 0 the tracking is perfect. C1) k should be as small as possible for precision Constraint 2 – tracking should be fast However, a smaller value of k leads to more iterations and thus a slower algorithm. C2) k should be as large as possible for speed Constraint 3 – tracking should stick to positive scales The new scale (ψ ′ ) is only guaranteed to be non-negative if: C3) 0 ≤ k ≤ 1. Constraint 4 – HSM-inspired Define the volume of a kernel to be: V ( s ) := ( 2π s ) 2 so that the derivative of this volume with N

N V (s) 2s Next, to make use of an idea from the HSM algorithm, we define the weight of points observed by a kernel (analogous to the number of samples in a histogram bin) to be: W ( x , s ) := V ( s ) D ( x , s ) and the log-transformed weight to be: W ∗ := ln W = ln V + D∗

respect to scale is: V ′ ( s ) =

We want to calculate how much the weight of points observed by a kernel can change as we shift along the log density gradient vector. For this purpose we calculate the derivatives of the log-weight. As in the previous section we calculate these derivatives in the local coordinate system and evaluate them at the origin of the system: d ∗ d d ∗ = ln V (ψ ) + D ∗ x h ,ψ = W x h ,ψ D x h ,ψ dh dh dh h =0 h =0 h =0 1 d ds d d d ∗ = ln V ( s ) + = ( Nk 2 ) 2 + W ∗ ( χ , s (m )) D∗ ( χ , s ( m ) ) D ( χ , s (m )) dm dm ds dm dm m =0 s =ψ m =0 m=0

( ( ) ))

(

(() )

(() )

To estimate the change in the log-weight due to a shift by the log-density gradient vector we linearly approximate the log-weight using the derivatives computed above. Thus: ⎛⎛ d ⎞ ⎞ ⎛⎛ d ⎞ ⎞ ∆W ∗ ≈ ⎜ ⎜ dhd ⎟ D ∗ ⎟ ⋅ ⎜ ⎜ dhd ⎟ W ∗ ⎟ ⎝ ⎝ dm ⎠ ⎠ ⎝ ⎝ dm ⎠ ⎠ h =0, m =0

=

d dh

2

D∗ +

d dm

d D ∗ + ( Nk 2 ) 2 ( dm D∗ ) 2

1

h =0, m =0

= ψ −1 C ( χ ,ψ ) − χ + 12 N k ψ −1ω ( χ ,ψ ) (ω ( χ ,ψ )ψ −1 − 1) 2

This change will be as negative as possible if there is no position change (i.e. C ( χ ,ψ ) = χ ), and the scale of the weighted samples is half that of the kernel scale (i.e. ω ( χ ,ψ ) = 12 ψ ). Hence we can assert the following (approximate because of the linearization step) bound on the change of Nk the log-weight: ∆W ∗ ≥ − 8

Taking a cue from the success of the HSM algorithm we can require that the weight under the aperture never reduces quicker than by 50% i.e. ∆W ∗ ≥ − ln 2 . If we combine this constraint with the bound derived above we obtain our fourth constraint: C4) k ≤ N −1 ln 256 . Constraint 5 – HRM-inspired Taking a cue from the success of the HRM algorithm we can require that the width of the kernels never reduces quicker than by 50%. This means that the scale never reduces to less than 25%, which gives us: k ≤ 43 C5) Combined constraints Combining all five constraints, plus adding a degree of preference for precision over speed, we arrive at our recommendation for k: k = min ( 12 , N −1 ln 256 ) .

This means that k only drops from

1 2

for N ≥ 12 .

2.3 Implementation Details The MSMS algorithm is simple to implement. Given n N-dimensional sample points pi . First a starting location and scale is chosen. Then the location and scale are repeatedly shifted until a stopping criteria is fulfilled. If the current location and scale are χ and ψ respectively, then the shifts are based on the kernel-induced weights for each sample calculated using the equation p− χ 2ψ

2

−1

⎛ ⎞ wi = e . The new location is given by χ ′ = ⎜ ∑ wi ⎟ ∑ wi pi . The new scale is computed as ⎝ i ⎠ i ψ ′ = (1 − k )ψ + kω ; where ω is the scale of the points under the kernel computed by −

−1

⎛ ⎞ 2 ω = ⎜ N ∑ wi ⎟ ∑ wi pi − χ , and the weight k is given by k = min ( 12 , N −1 ln 256 ) . i ⎝ ⎠ i For the starting location we use the mean of the sample points (i.e. χ 0 := n −1 ∑ pi ). For i

the starting scale we use a value sufficiently large so that the weight of points seen by the kernel is at least 95% of the total i.e.

∑e



p − χ0 2ψ 0

2

≥ 0.95n . We find this value by initializing

i

ψ 0 := N −1 ∑ pi − χ , and then repeatedly doubling ψ 0 until the 95% criteria is fulfilled. 2

i

Ideally, the algorithm would be allowed to keep on tracking until only a single sample is visible within the kernel, but numerical inaccuracies sometimes result in an oscillating behavior of extremely small shifts when just two, three or four samples are visible within the kernel. Instead, to robustly terminate the shifting process we stop when the magnitude of the location and scale shifts is sufficiently small compared to the current scale. We have found the criteria

( ∆χ

2

)

+ ( ∆ψ ) s ≤ 10−3ψ effective. 2

A final optional detail that we find more than doubles the speed of the algorithm without any appreciable reduction in performance is to gradually inactivate sample points as they lie too 2 pi − χ far away from current kernel. Formally, if at any stage > 709 then the sample point pi 2ψ is ignored for the remainder. The threshold 709 is such that the weight of the point is zero at 32bit floating point precision. When using this deactivation strategy we stop the shifting process if only two or fewer samples are active, or if the small-shift criteria is met. In practice, termination on the basis of the number of active points is far more common.

3. Performance Characterization In this section we evaluate the performance of the MSMS algorithm and compare it to the stateof-the-art. Following our discussion in section 1.4, for 1-D datasets the algorithms we make comparison with are the Half Range Mode (HRM), the Robust Parametric Mode (RPM), and Gasser’s version of the Bootstrap approach (GBOOT). Of these algorithms, only GBOOT has a multi-dimensional version, and we compare MSMS to some published performance figures for this. Finally, we demonstrate MSMS applied to the very high-dimensional problem that originally motivated us to develop the algorithm.

3.1 1-D datasets We have selected a test suite of six 1-D distributions upon which to assess mode estimation performance. Some of these distributions resemble ones used elsewhere in the mode estimation literature, but the suite as a whole has been put together by us to provide a wide range of tests. The distributions, shown in figure 5, are a Normal ( σ = 1 ), a log-Normal ( µ and σ parameters both 1), a Pareto (parameters, k = 1 and α = 12 ), a bi-modal distribution (the sum of two Normals, 0.4 N ( −1,0.6 ) + 0.6 N (1,0.6 ) , with mode=0.995), a platykurtotic distribution (the sum

of two Normals, 0.4 N ( −0.75,0.8 ) + 0.6 N ( 0.75, 0.8 ) , with mode=0.486), and a spike distribution (the sum of a uniform and a normal distribution 0.87).

19 20

χ[ −1.74,1.74] + 201 N ( −0.87, 0.09 ) , with mode = –

spike Pareto platykurtotic

Normal

bi-modal

−4

log-Normal

−2

0

2

4

Figure 5 – Shows the six 1-D distributions used for comparing performance of mode estimators.

The comparative performance of the mode estimators on the test suite are shown in figures 6 and 7. The figures show the root-mean-squared (RMS) estimation error and the bias (absolute value of the mean signed-estimation-error). For the MSMS, RPM and HRM algorithms, sets of samples of cardinality between 25 and 211 were used, and 1000 trials were made for each size. For GBOOT, the maximum cardinality assessed was only 29 and only 100 trials at each size were performed. This was because the computation time for this algorithm is considerably longer than the others, and the partial results were sufficient to determine the general pattern.

3

RPM GBOOT HRM MSMS

0.8

0.4 2 RMS error

0.4

0.2

1 25 26 27 28 29 210 211

25 26 27 28 29 210 211 3

bias

25 26 27 28 29 210 211 0.8

0.4 2 0.4

0.2

1 25 26 27 28 29 210 211

Normal

25 26 27 28 29 210 211

log-Normal

25 26 27 28 29 210 211

Pareto

Figure 6 – Comparative performance of the tested mode estimators on three of the test distributions shown in figure 5. Note that the curves for the GBOOT algorithm applied to the Pareto distribution are not shown as they are off scale.

Considering the results shown in figure 6, we observe that in all three cases the performance of the HRM and MSMS algorithms are very similar in RMS error and bias. GBOOT performs better than these two algorithms for unskewed distributions (i.e. the Normal) but worse for moderate skew (i.e. the log-Normal) and much worse for high skew (i.e. the Pareto). In fact, the RMS error for the GBOOT algorithm operating on samples from the Pareto distribution is so large (rising from 103 to 106) that it has not been plotted in the figure. Best, on balance, are the RPM results.

RMS error

RPM GBOOT HRM MSMS

0.8

0.6

1

0.4

0.3

0.5

25 26 27 28 29 210 211

25 26 27 28 29 210 211

25 26 27 28 29 210 211

0.8

0.6

1

0.4

0.3

0.5

bias

25 26 27 28 29 210 211

25 26 27 28 29 210 211

bi-modal

25 26 27 28 29 210 211

platykurtotic

spike

Figure 7 – Comparative performance of the tested mode estimators on three of the test distributions shown in figure 5.

Considering the results shown in figure 7, we observe that the HRM and MSMS algorithms also perform similarly for these distributions. GBOOT performs less well than these two except at small sample size. Again, the RPM algorithm performs well compared to the HSM and MSMS, but with two caveats: first is that although the RMS error is often smaller, the bias is generally greater, second is that it is clearly out-performed in the case of large sample sets from the spike distribution. RPM GBOOT HRM MSMS

1

1

0.8

0.8

0.6

0.6

0.4

0.4

0.2

0.2 25

26

27

28

RMS error

29

210

211

25

26

27

28

29

210

211

bias

Figure 8 – Shows the performance scores of figures 6 and 7 averaged across the six test distributions. The Pareto scores for the GBOOT algorithm were left out of this averaging.

Figure 8 shows the performance scores averaged across all six test distributions, except for the GBOOT algorithm for which the Pareto scores were left out. This averaging across distributions

is an artificial operation, but because the distributions were chosen to have similar supports it gives quite an informative overview of performance. It shows that RPM is the best performing algorithm for smaller sample sizes, but as the sample size increases HRM and MSMS beat it first in bias and then later also match it in RMS error. The graph confirms that there is nothing to choose between HRM and MSMS in terms of bias, but that the MSMS has a very slight edge in RMS error. The GBOOT algorithm performs the worst, even with the Pareto data left out of the pooling: the performance is still the worst if the log-Normal data is also excluded, but the margin is smaller.

3.2 Multi-dimensional datasets An important strength of the MSMS algorithm is that it operates on multi-dimensional data as readily as one-dimensional. This is not true of the RPM, HRM or HSM algorithms. No multidimensional extensions of these algorithms have been published and the only such extensions that we can conceive of are complex and likely to be very slow. However, the GBOOT algorithm does naturally extend to multiple dimensions, and performance figures for a particular multidimensional distribution have been published [8]. The distribution used was a separable product of 20 chi-squared distributions (ν = 8 ), and sample sets of size 50 and 200 were evaluated. In Table 1 we give the performance figures for mode estimation based on GBOOT and MSMS. The figures for GBOOT are computed from the figures in the original paper, not from our implementation. Thus it can be expected that the aspects of GBOOT which require hand-tuning (i.e. scales to try and number of bootstraps to perform) are well-chosen.

50 200

GBOOT 2.08 (1.92) 1.66 (1.60)

MSMS 2.68 (1.27) 2.14 (1.03)

Table 1 – RMS errors and biases (in brackets) of mode estimators applied to the multidimensional distribution specified in [8].

Table 1 shows that, on the test distribution, the GBOOT RMS errors are about 30% less than those of the MSMS, but the biases are 50% larger. These results are consistent with those of the 1-D study. Recall that figure 6 showed that GBOOT performed well with unskewed distributions, but the performance rapidly became poor as skew increased. The chi-squared distributions used in this multi-dimensional study have a skewness of 1.0, to be compared with the skewness of 0.0 of the Normal, and the skewness of 6.2 of the log-Normal distributions used in the 1-D study. Thus it is plausible that the chi-squared distributions are still in the low-skew range in which GBOOT performs well. Although we judge that on balance the 1-D and n-D results that we have presented make the case for the superiority of the MSMS algorithm, the picture is not completely clear cut. Further investigation, particular of the n-D case is desirable, but is difficult to do given the differences in computation speed between the various algorithms. In particular, in our implementations, we find the GBOOT algorithm to be 103-104 slower than the MSMS.

The speed of the MSMS algorithm allows it to be applied to problems that the GBOOT algorithm could not. In particular, the problem that originally we developed MSMS for involves computing mode estimates in a space of 4,096 dimensions based on 350,000 samples. The aim of the computation, whose motivation is described elsewhere [43-45], was to discover the modal 64µ64 patch from a database of natural images. The patches were selected at random, but were rotated and intensity-scaled so that their 0th and 1st order structures (as measured by centrallylocated Gaussian derivative filters of scale σ ≈ 7 ) agreed. Figure 9 shows that the most likely patch in this circumstance is a straight step edge. We have shown elsewhere [44, 45] that if Gaussian or Brownian noise images are used instead of natural, the results are different and agree with theoretical expectation, confirming the correct operation of the MSMS algorithm even in very high-dimensional spaces.

Figure 9 – On the left is an estimate of the modal natural image patch, conditioned on equal 0th and 1st order structure, computed using the MSMS algorithm. The patch is 64µ64 pixels, and the estimate was based on 350,000 patches. On the right is shown the patch as it appears in the image from which it originated.

4. Discussion In the scale space approach to image analysis it is normal to distinguish between superficial structure that exists as some particular scale of analysis and deep structure which is the unfolding of superficial structure as scale changes from coarse to fine [36]. There are two issues in deep structure that are relevant to any mode-estimation algorithm that makes use of kernel-based density-estimation and we discuss them in the following sub-sections, paying particular notice of how the MSMS algorithm deals with them.

4.1 Multi-Modality If a distribution has multiple local maxima then it is said to be multi-modal. In such a case, we require a criterion for choosing one of the modes as the most representative of the distribution. One strategy is to choose the highest mode (the ‘global mode’), but this is not the only reasonable possibility. An alternative choice, the ‘stable mode’, is based on the behavior of a distribution as it is blurred. If the blurring is performed with Gaussian kernels, then the modes

can change in position continuously, but for 1-D distributions they are never created, only destroyed, with increasing scale. Eventually, for sufficiently coarse blurring, only a single mode survives. Since modes in 1-D are never created by blurring, this surviving mode must correspond to one of the modes of the original unblurred distribution, this is the ‘stable mode’ [46]. Informally, the stable mode tends to be the mode that is associated with the most weight of distribution, and this need not coincide with the global mode.

100%

50%

25

28

211

25

28

211

25

28

211

100%

50%

100%

50%

Figure 10 – In the left column are three very similar distributions. The central column shows the scale space plots of these distributions, each row of a density plot corresponding to the distribution blurred to a different degree. Blurring increases exponentially from bottom to top of the density plots. The curves overlaying the density plots show the exact tracks of the modes as they vary in position as blur is increased. It can be seen that the small quantitative differences in the original distributions lead to qualitative differences in which of the three fine scale modes is the stable mode. The plots at the right illustrate results of applying the MSMS algorithm to sets of samples from the distributions at left. The sizes of the sets of samples range from 24 to 211. 1000 trials were run at each sample size. The black band indicates what fraction of the MSMS estimates were closest to the leftmost mode, the grey band to the central mode and the white band to the rightmost mode. It can be seen that for all three distributions, as the sample size is increased, the stable mode is correctly identified in the majority of trials. The black and white curves overlaid on the right-hand plots show the band-boundaries if the MSMS algorithm employs the more cautious scale reduction factor k = 1 4 rather than the recommended default k = 1 2 used elsewhere in the paper. The fact that the boundaries are essentially unchanged suggests that the recommended value is sufficiently small.

The MSMS algorithm, working as it does by tracking across scale, is designed to discover the stable not the global mode. In figure 10 we present results showing that it is successful at this aim, though naturally its rate of success depends upon the distribution and the number of samples from it. Figure 10 also shows that the stable mode, like the global mode, is precarious in the sense that small changes to the distribution can result in it changing discontinuously. Figure 10 also presents evidence that our recommended value of 12 for the scale-reduction weight k (see section 2.2) is sufficiently small for good cross-scale tracking.

4.2 Mode Creation The deep structure of the modes of a multi-dimensional distribution is less constrained than for a one-dimensional. In particular, modes can be created by blurring even when a Gaussian kernel is used [36]. Figure 11 shows an example 2-D distribution for which a mode is created by blurring, and for which the mode that it is created actually goes on to be the one that survives to arbitrarily coarse scales. In theory and in practice, the MSMS algorithm when applied to such a distribution tracks down to the scale-location where the surviving mode is created and then terminates because the computed shifts are of negligible size. This termination of the algorithm at a mode creation is readily detectable as it is signaled by the number of active points seen by the final kernel being more than just one or two.

Figure 11 – Illustrates that for multi-dimensional distributions, modes can be created by blurring. At left, the 2-D distribution shown as a height plot has only a single mode when unblurred. At right, the same distribution is shown as a contour plot on the ground plane. Above it are shown the trajectories of the modes as scale increases vertically. After a small amount of blur the narrow ridge is dissipated causing a new mode to appear. Further blurring causes the original mode to be destroyed, leaving the created mode as the one that survives to arbitrarily coarse scales.

5. Conclusion We have presented a novel algorithm – the Multiscale Mean Shift (MSMS) – for estimating the mode of a distribution on the basis of samples from it. The algorithm developed from our earlier work on mode estimation by Scale Space tracking, but supersedes that approach. We have explained how the algorithm is distinct from but related to the Mean Shift and Half Range Mode algorithms presented by other authors. In fact the MSMS algorithm can be regarded as a combination of these two approaches. We have explained how the MSMS algorithm arises from the kernel-based density-estimation approach, in the particular case that the kernels employed are Gaussians. We have shown that for Gaussian kernels alone, the mode of the estimated density at different scales can be tracked by ascending the density across location and scale. We have derived a tracking scheme based on a sequence of shifts of location and scale, where the shifts have two interpretations. Either as the gradient vector of the log density when expressed in a suitable local coordinate scheme, or as a simple operation on the samples weighted by the kernel of the current location-scale. We note that the identity between these two methods of arriving at the MSMS scheme relies on two ‘tricks’. The first – consideration of the logged rather than the raw density – is also standard to mean shift derivations. The second – the use of local coordinate schemes – is an alternative to the use of a scale-dependent, gradient-vector multiplier as in standard in mean shift derivations. We use the choice of local coordinate system to produce not only the desired location-shift multiplier, but also the desired scale-shift multiplier. Neither ‘trick’ has been derived in any sense, rather our derivation follows from them. Thus, at present, their justification comes only from the empirically-assessed effectiveness of the MSMS algorithm We have compared the performance of the MSMS algorithm when applied to 1-D distributions to state-of-the-art algorithms, and found it competitive in RMS error, bias-minimization and computational speed. We have performed a limited comparison with the only state-of-the-art algorithm that is able to operate on multi-dimensional data and found it competitive in RMS error and superior in bias-minimization. A more extensive comparison was not possible because of the very slow speed of the only rival algorithm. We have shown that the fast speed of our algorithm allows us to perform mode estimation for high-dimensional problems that no other algorithm would be able to tackle. We have referred to work that has validated the results of the algorithm when applied to these very high-dimensional problems. The MSMS algorithm has only one parameter – the scale reduction weight k - that our derivation failed to determine precisely. We have presented arguments that allowed us to make a recommendation for a value of k that should be general purpose and not require problem-specific tuning. We have presented data that support the value of k that we have recommended. We have discussed the deep structure of modes as they vary with the scale used in density estimation and have noted two issues. First, that the mode tracked down to may depend precariously on the distribution (just as the global mode of a bi-modal distribution with two nearequal height peaks can). Second, that for multi-dimensional distributions, there can be modes created by blurring that prevent tracking down to the unblurred distribution. To our mind, both of these issues are arguments in favor of the full determination of the possibly several modes of a distribution, as being the only truly robust approach to mode analysis. The current version of the

MSMS algorithm does not detect multiple modes, but this is an issue that we hope to address in future work.

References 1. Bickel, D.R., Robust estimators of the mode and skewness of continuous data. Computational Statistics and Data Analysis, 2002. 39: p. 153-163. 2. Baxter, M.J., C.C. Beardah, and R.V.S. Wright, Some archaeological applications of kernel density estimates. Journal of Archaeological Science, 1997. 24(4): p. 347-354. 3. Marchette, D.J. and E.J. Wegman, The filtered mode tree. Journal of Computational and Graphical Statistics, 1997. 6(2): p. 143-159. 4. Hedges, S.B. and P. Shah, Comparison of mode estimation methods and application in molecular clock analysis. Bmc Bioinformatics, 2003. 4: p. art. no.-31. 5. Ooi, H., Density visualization and mode hunting using trees. J. Comp. & Graph. Statist., 2002. 11(2): p. 328-347. 6. Comaniciu, D., An algorithm for data-driven bandwidth selection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2003. 25(2): p. 281-288. 7. Comaniciu, D. and P. Meer, Mean Shift: A robust approach toward feature space analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2002. 24(5): p. 603-619. 8. Gasser, T., P. Hall, and B. Presnell, Nonparametric estimation of the mode of a distribution of random curves. Journal of the Royal Statistical Society Series B-Statistical Methodology, 1998. 60: p. 681-691. 9. Georgescu, B., I. Shimshoni, and P. Meer, Mean shift based clustering in high dimensions: A texture classification example, in Ninth Ieee International Conference on Computer Vision, Vols I and Ii, Proceedings. 2003. p. 456-463. 10. Liu, X. and H.-G. Muller, Modes and clustering for time-warped gene expression profile data. Bioinformatics, 2003. 19(15): p. 1937-1944. 11. Hall, P. and N.E. Heckman, Estimating and depicting the structure of a distribution of random functions. Biometrika, 2002. 89(1): p. 145-158. 12. Cousineau, D., S. Brown, and A. Heathcote, Fitting distributions using maximum likelihood: methods and packages. Behavior Research Methods, Instruments and Computers, 2004. 36(4): p. 742-756. 13. Everitt, B.S. and D.J. Hand, Finite Mixture Distributions. 1981, London: Chapman Hall. 14. Bickel, D.R., Robust and efficient estimation of the mode of continuous data: The mode as a viable measure of central tendency. Journal of Statistical Computation and Simulation, 2003. 73(12): p. 899-912. 15. Rosenblatt, M., Remarks on some non-parametric estimates of a density function. Ann. Math. Stats., 1956. 27: p. 642-669. 16. Parzen, E., On estimation of probability density function and mode. Annals of Mathematical Statistics, 1962. 33: p. 520-531. 17. Silverman, B.W., Density Estimation for Statistics and Data Analysis. 1986, London: Chapman Hall. 18. Fukunaga, K. and L.D. Hostetler, The estimation of the gradient of aq density function, with applications in pattern recognition. IEEE Transactions on Information Theory, 1975. 21: p. 32-40.

19. Cheng, Y., Mean Shift, Mode Seeking, and Clustering. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1995. 17(8): p. 790-799. 20. Wand, M.P. and M.C. Jones, Kernel Smoothing. 1995, London: Chapman Hall. 21. Fukunaga, K., Statistical Pattern Recognition. 2nd ed. 1990, New York: Academic Press. 22. Scott, D.W., R.A. Tapia, and J.R. Thompson, Kernel density estimation revisited. Nonlinear Analysis, Theory, Methof & Application, 1977. 1: p. 339-372. 23. Hazelton, M.L., An optimal local bandwidth selector for kernel density estimation. Journal of Statistical Planning and Inference, 1999. 77(1): p. 37-50. 24. Katkovnik, V. and I. Shmulevich, Kernel density estimation with adaptive varying window size. Pattern Recognition Letters, 2002. 23(14): p. 1641-1648. 25. Efron, B. and R. Tibshirani, An Introduction to the Bootstrap. 1993, New York, London: Chapman and Hall. 26. Romano, J.P., Bootstraping the mode. Annals of the Institute of Statistical Mathematics, 1988. 40: p. 565-586. 27. Faraway, J.J. and M. Jhun, Boostrap choice of bandwidth for density estimation. J. Am. Statist. Ass., 1990. 85: p. 1119-1122. 28. Grund, B. and P. Hall, On the minimisation of L(p) error in mode estimation. Annals of Statistics, 1995. 23(6): p. 2264-2284. 29. Wang, S.J., On the Bootstrap and Smoothed Bootstrap. Communications in Statistics-Theory and Methods, 1989. 18(11): p. 3949-3962. 30. Griffin, L.D. and M. Lillholm. Mode Estimation by Pessimistic Scale Space Tracking. in Scale Space '03. 2003. Isle of Skye, UK: Springer. 31. Minnotte, M.C., D.J. Marchette, and E.J. Wegman, The bumpy road to the mode forest. Journal of Computational and Graphical Statistics, 1998. 7(2): p. 239-251. 32. Minnotte, M.C. and D.W. Scott, The Mode Tree: A Tool for visualization of nonparametric density features. J. Comp. & Graph. Statist., 1993. 2: p. 51-68. 33. Silverman, B.W., Using Kernel density Estimates to investigate Multimodality. J. Roy. Statist. Soc. B, 1981. 43: p. 97-99. 34. Koenderink, J.J., The Structure of Images. Biological Cybernetics, 1984. 50(5): p. 363-370. 35. Griffin, L.D., Descriptions of Image Structure. 1995, London: PhD thesis, University of London. 36. Griffin, L.D. and A.C.F. Colchester, Superficial and Deep-Structure in Linear Diffusion Scale-Space - Isophotes, Critical-Points and Separatrices. Image and Vision Computing, 1995. 13(7): p. 543-557. 37. Scale Space '99. in Scale Space '99. 1999. Corfu, Greece: Springer. 38. Scale Space '01. in Scale Space '01. 2001. Vancouver, Canada: Springer. 39. Scale Space '03. in Scale Space '03. 2003. Isle of Skye, UK.: Springer. 40. Scale Space '05. in Scale Space '05. 2005. Hofgeismar, Germany: Springer. 41. Weickert, J., S. Ishikawa, and A. Imiya, On the history of Gaussian scale-space axiomatics, in Gaussian Scale-Space Theory. 1997. p. 45-59. 42. Press, W.H., et al., Numerical Recipes in C++. 2nd ed. 2002, Cambridge: Cambridge University Press. 43. Griffin, L.D. and M. Lillholm, Image features and the 1-D, 2nd order gaussian derivative jet, in Proc. Scale Space 2005. 2005, Springer. p. 26-37. 44. Griffin, L.D., M. Lillholm, and M. Nielsen, Natural image profiles are most likely to be step edges. Vision Research, 2004. 44(4): p. 407-421.

45. Griffin, L.D., Feature classes for 1-D, 2nd order image structure arise from the maximum likelihood statistics of natural images. Network-Computation in Neural Systems, 2005. in press. 46. Griffin, L.D., Scale-imprecision space. Image and Vision Computing, 1997. 15(5): p. 369398.

A Multiscale Mean Shift Algorithm for Mode Estimation ...

Computer Science, University College London, UK; [email protected] ... algorithm are that (i) it is simple, (ii) it is fast, (iii) it lacks data-dependent parameters ...

723KB Sizes 1 Downloads 296 Views

Recommend Documents

A Multiscale Mean Shift Algorithm for Mode Estimation 1. Introduction
Computer Science, University College London, UK; [email protected]. 2 ...... 36(4): p. 742-756. 13. Everitt, B.S. and D.J. Hand, Finite Mixture Distributions.

Stable Mean-Shift Algorithm And Its Application To The ieee.pdf ...
Stable Mean-Shift Algorithm And Its Application To The ieee.pdf. Stable Mean-Shift Algorithm And Its Application To The ieee.pdf. Open. Extract. Open with.

A Fast Algorithm For Rate Optimized Motion Estimation
Abstract. Motion estimation is known to be the main bottleneck in real-time encoding applications, and the search for an effective motion estimation algorithm has ...

A Fast Algorithm For Rate Optimized Motion Estimation
uous motion field reduces the bit rate for differentially encoded motion vectors. Our motion ... In [3], we propose a rate-optimized motion estimation based on a “true” motion tracker. ..... ftp://bonde.nta.no/pub/tmn/software/, June 1996. 477.

Mean-shift and hierarchical clustering for textured ...
Sci. & Software Eng. Dept., Laval Univ., Quebec City, Que., Canada . Touzi, R.‪‬ ... collective works, for resale or redistribution to servers or lists, or reuse of any.

On Weight Ratio Estimation for Covariate Shift - Carnegie Mellon ...
Covariate shift is a common assumption for various Domain Adaptation (DA) tasks. Many of the .... Weight ratio A common way of restricting the divergence of source and target weights marginal .... In Proceedings of the Conference on Algorithmic Learn

Add shift to -
Aug 19, 2017 - This paper proposes adding shift algorithms to the C++ STL which shift elements forward or backward in a range of elements. II. Motivation and ...

Add shift to -
not cycle with the data, the mask indices need to be updated whenever the buffer is cycled. - A programmer might need to shift elements in non-circular buffers ...

DOA Estimation Using MUSIC Algorithm for Quantized ...
confined to the single parameter per source case (e.g. azimuth angle) only. However, the analysis presented can be easily extended to the multiple pa- rameter case. Let us consider M radiating narrowband sources with center frequency fo observed by a

Mean, median and mode filtering of images
Medical Imaging Science Interdisciplinary Research Group,. King's College, London, UK ([email protected]) ..... that seem artefactual, for example the dark spur extending from the top left of the ... Lecture Notes in Computer Science, vol.

Mean, Median, Mode, and Range Color by Numbers.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Mean, Median ...

A MULTISCALE APPROACH
Aug 3, 2006 - ena (e.g. competition for space, progress through the cell cycle, natural cell death and drug-induced cell .... the host tissue stim- ulates the production of enzymes that digest it, liberating space into which the .... The dynamics of

Object tracking using SIFT features and mean shift
Jul 25, 2012 - How scale-invariant feature transform. (SIFT) works. • Algorithm, Ref. [3]:. – Keypoint localization. • Interpolation of nearby data for accurate position. • Discarding low-contrast keypoints. • Eliminating edge responses. â€

Robust Graph Mode Seeking by Graph Shift
ample, in World Wide Web, dense subgraphs might be communities or link spam; in telephone call graph, dense subgraphs might be groups of friends or families. In these situations, the graphs are usually very sparse in global, but have many dense subgr

Fast Mean Shift with Accurate and Stable Convergence
College of Computing. Georgia Institute of Technology. Atlanta, GA 30332 ... puter vision community has utilized MS for (1) its clus- tering property in .... rameter tuning for good performance.2 ...... IEEE Trans. on Information Theory,. 21, 32–40

Agglomerative Mean-Shift Clustering via Query Set Compression ∗
To find the clusters of a data set sampled from a certain unknown distribution is important in many machine learning and data mining applications. Probability.

Agglomerative Mean-Shift Clustering via Query Set ... - CiteSeerX
To find the clusters of a data set sampled from a certain unknown distribution is important in many machine learning and data mining applications. Probability.

Agglomerative Mean-Shift Clustering via Query Set ... - CiteSeerX
learning and data mining applications. Probability ..... Figure 1: Illustration of iterative query set compression working mechanism on a 2D toy dataset. See text for the ..... MS and LSH-MS, lies in that it is free of parameter tuning, hence is more

Mode Estimation using Pessimistic Scale Space Tracking
1 For some non-generic distributions, this definition fails to define a unique ..... mode estimation method used is pessimistic tracking with 3.78 sds of pessimism.

Fast Sub-Pixel Motion Estimation and Mode Decision ...
partition selection and only performs the 'precise' sub-pel search for the best ... Motion Estimation process contains two stages: integer pixel search .... full. COST avg. COST. COST avg. COST ii blocktype if c where COSTFull is the best COST after

A Fast Sub-Pixel Motion Estimation Algorithm for H.264/AVC Video ...
H.264/AVC is the state-of-the-art video coding standard ... ommended by Associate Editor R. Lukac. .... where R(MV) is the number of bits to code the MV and.

Noise-contrastive estimation: A new estimation principle for ...
Any solution ˆα to this estimation problem must yield a properly ... tion problem.1 In principle, the constraint can always be fulfilled by .... Gaussian distribution for the contrastive noise. In practice, the amount .... the system to learn much

Adaptive Least Mean Squares Estimation of Graph ...
processing tools to signals defined over a discrete domain whose elementary ... tated by the graph topology, the analysis tools come to depend on the graph ...... simulations. D. Sampling Strategies. As illustrated in the previous sections, the prope