Stat Comput (2008) 18: 15–26 DOI 10.1007/s11222-007-9034-y

Accuracy of edge detection methods with local information in speckled imagery Juliana Gambini · Marta E. Mejail · Julio Jacobo-Berlles · Alejandro C. Frery

Received: 17 August 2006 / Accepted: 19 July 2007 / Published online: 11 September 2007 © Springer Science+Business Media, LLC 2007

Abstract We compare the accuracy of five approaches for contour detection in speckled imagery. Some of these methods take advantage of the statistical properties of speckled data, and all of them employ active contours using B-spline curves. Images obtained with coherent illumination are affected by a noise called speckle, which is inherent to the imaging process. These data have been statistically modeled by a multiplicative model using the G0 distribution, under which regions with different degrees of roughness can be characterized by the value of a parameter. We use this information to find boundaries between regions with different textures. We propose and compare five strategies for boundary detection: three based on the data (maximum discontinuity on raw data, fractal dimension and maximum likelihood) and two based on estimates of the roughness parameter (maximum discontinuity and anisotropic smoothed roughness estimates). In order to compare these strategies, a Monte Carlo experience was performed to assess the accuracy of fitting a curve to a region. The probability of finding

J. Gambini · M.E. Mejail · J. Jacobo-Berlles Facultad de Ciencias Exactas y Naturales, Universidad de Buenos Aires, Ciudad Universitaria, Pabellón I, C1428EGA Buenos Aires, Argentina J. Gambini e-mail: [email protected] M.E. Mejail e-mail: [email protected] J. Jacobo-Berlles e-mail: [email protected] A.C. Frery () Instituto de Computação, Universidade Federal de Alagoas, BR 104 Norte km 97, 57072-970 Maceió, AL, Brazil e-mail: [email protected]

the correct edge with less than a specified error is estimated and used to compare the techniques. The two best procedures are then compared in terms of their computational cost and, finally, we show that the maximum likelihood approach on the raw data using the G0 law is the best technique. Keywords Active contours · B-spline curve fitting · Image analysis · SAR imagery · Speckle noise 1 Introduction Several types of imaging devices employ coherent illumination as, for instance, Synthetic Aperture Radar (SAR), sonar, laser and ultrasound-B. These images are affected by a noise called speckle, a kind of degradation that does not obey the classical hypotheses of being Gaussian and additive. Speckle noise reduces the ability to extract information from the data, so specialized techniques are required to deal with such imagery. Identifying boundaries that separate different areas is one of the most important image understanding procedures (for a medical application see, for instance, the work by Figueiredo and Leitão 1992). High level image processing relies, among other features, on precise and accurate boundaries. Finding boundaries between regions of different roughness is a hard task when data is contaminated by speckle noise. In this work we present a comparative study of five approaches for contour detection using local information in speckled imagery, three of them based on the statistical properties of data and B-spline curves and one that uses fractal dimension. Four of these proposals, namely all but discontinuity on raw data, are entirely new (maximum likelihood and discontinuity on roughness) or novel applications of known techniques (fractal dimension on speckled data and anisotropic diffusion on the roughness).

16

The B-spline approach has been widely used in curve representation for boundary detection (Cipolla and Blake 1990), shape approximation (Medioni and Yasumoto 1986) and object tracking (Brigger et al. 2000). Contours formulated by means of B-splines allow for local control, have local representation, require few parameters and are intrinsically smooth. Speckled data have been statistically modeled under the multiplicative model using the family of G distributions, since these probability laws are able to describe the observed data better than other laws, specially in the case of rough and extremely rough areas. Such situations are common when scanning urban spots or forests on undulated relief, and for them the more classical  and K distributions do no exhibit good performance (Frery et al. 1997; Mejail et al. 2001). Under the G model, regions with different degrees of roughness can be characterized by the parameters. Therefore, this information can be used to find boundaries between regions with different textures (see, for instance, Gambini et al. 2006). We propose and compare five strategies for boundary detection: three that directly employ the image values (maximum discontinuity on raw data, fractal dimension and maximum likelihood under the GA0 model) and two based on estimates of the roughness parameter (maximum discontinuity and anisotropic smoothed roughness estimates). A first naïve approach for boundary detection would be scanning through candidate directions searching for discontinuity points. But, as will be presented, edge detection in speckled imagery is difficult due to the low signal-to-noise ratio. In order to alleviate this issue, one can extract meaningful features from the raw data as, for instance, the fractal dimension. This approach has been used for texture and roughness based image segmentation (see, for instance, Dennis and Dessipris 1989; Keller 1989; Liu and Li 1997; Peli 1990); in particular, Blackledge and Fowler (1992) and Du and Yeo (2002) use estimates of the fractal dimension in SAR image understanding. The second algorithm we assess consists of estimating the fractal dimension using the box-counting method. As pointed by, among others, Oliver (1991), statistical features can be successfully used with this kind of data: Germain and Réfrégier (2001a, 2001b) use parameter estimation of the Gamma distribution for edge location and segmentation, while Mejail et al. (2003) use the estimated parameters of the GA0 law for SAR image classification. In this paper the estimated parameters of the GA0 distribution are extracted as features for edge detection, either for seeking contours directly in these estimates or after smoothing the estimates using anisotropic diffusion. The five methods under assessment will be compared estimating the probability of finding an edge with an error of

Stat Comput (2008) 18: 15–26

less than a specified number of pixels. This assessment will be made using a Monte Carlo experience with simulated images. The robustness of the best procedures with respect to the starting solution is commented and, finally, an example using real data on scenes with varying complexity is presented. The contribution of this paper is threefold, namely (i) the proposal of edge detection procedures that employ the GA0 law, (ii) the use of five edge detection techniques for speckled imagery: three based on the data (maximum discontinuity on raw data, fractal dimension and maximum likelihood) and two based on estimates of the roughness parameter (maximum discontinuity and anisotropic smoothed roughness estimates) and (iii) the assessment of these five techniques using a Monte Carlo setup; no quantitative results of this sort were found in the literature. We conclude that the maximum likelihood model assuming the GA0 distribution for the raw (speckled) data is the best of these five edge detection procedures with respect to both the error and the computational cost. This procedure can be used in any situation where statistical models are suitable for describing the data. The structure of this paper is as follows: Sect. 2 describes the statistical model used for single channel speckled data, Sect. 3 provides a brief account of B-spline curve fitting, Sect. 4 describes the algorithms in detail, Sect. 5 presents the error evaluation methodology, Sect. 6 presents the results and Sect. 7 concludes the paper.

2 The G distribution for speckled data Speckled images can be modeled as the product of two independent random fields, one corresponding to the backscatter X and other corresponding to the speckle noise Y (see Goodman 1976): Z = X · Y . For amplitude data, the speckle noise Y is modeled as a  −1/2 (n, n) distributed random variable, where n is the number of looks used to generate the image; this parameter is known or estimated beforehand, and it is valid for the whole image. The most general model for the backscatter X here considered is the square root of the Generalized Inverse Gaussian law (see Barndorff-Nielsen and Blæsild 1983; Jørgensen 1982; Seshadri 1993), which has the square root of Gamma and its reciprocal distributions as particular cases. These, in turn, give rise to the KA and the GA0 distributions for the return Z, respectively; see Frery et al. (1997) and Mejail et al. (2001). The G 0 distribution is an attractive choice for modelling speckled data, given its tractability, expressiveness and capability of retrieving detailed information from the data

Stat Comput (2008) 18: 15–26

17

(see, Quartulli and Datcu 2004, for its use in SAR imagery). Its density function for amplitude data, is given by fG 0 (z) = A

2nn (n − α) γ α (−α)(n)

−α, γ , z > 0,

z2n−1 , (γ + z2 n)n−α

n ≥ 1.

(1)

It is denoted Z ∼ GA0 (α, γ , n), and its moments are  EG 0 (Z r ) = A

γ 2n

r/2

 (−α + r/2)  (n + r/2) ,  (−α)  (n)

(2)

if −α > r or infinite otherwise. Given the data, the statistical parameters are estimated and this information is used to extract region boundaries present in the image. 2.1 Parameter interpretation One of the most important features of the GA0 distribution is that the parameter α has immediate interpretation in terms of roughness. For values of α near zero, the imaged area presents very heterogeneous gray values, as is the case of urban areas in SAR images. As we move to less heterogeneous areas like forests, the value of α diminishes, reaching its lowest values for homogeneous areas like pastures and certain types of crops. This is the reason why this parameter is regarded to as a roughness or texture measure. The parameter γ of the GA0 distribution is a scale parameter, that is, let W be a GA0 (α, γ , n) distributed random variable, then γ −1/2 W ∼ GA0 (α, 1, n); this property will be used to scale and threshold the data in Sect. 4.2. In this manner, this parameter is related to the amplitude of the backscatter. Our sight is more sensitive to variations in mean value than to those related to roughness and, therefore, it is easy to find situations were areas with different roughness parameters are perceived as alike, whilst areas with different scale parameters are immediately discriminated regardless of their roughness. Figure 1 shows simulated imagery under the GA0 law with a variety of parameters. It is noticeable that the bigger γ the brighter the image (from left to right this parameter is set to 10, 50, 100 and 1000). The roughness parameter has influence on the variability of the data and, therefore, its effect is less noticeable; in these images, α varies from very homogeneous targets (top, α = −12) to very heterogeneous ones (bottom, α = −1.5) with intermediate situations in between (α = −8 and −4). 2.2 Parameter estimation Several parameter estimation techniques are available, being the most remarkable ones those based on maximum likelihood and on the substitution method (see, for instance,

Fig. 1 Synthetic GA0 (α, γ , 1) imagery for γ ∈ {10, 50, 100, 1000} (left to right) and α ∈ {−1.5, −4, −8, −12} (bottom up)

Manski 1988). Among the procedures derived from the last, sample moments (MO for short) and order statistics are frequently used. To estimate α and γ it is necessary to compute two sample moments, and we chose to work with moments of order 1/2 and 1, namely m1/2 and m1 respectively. From (2), these moments are given by  1/4 (−α − 1/4)(n + 1/4) γ , m1/2 = n (−α)(n)  1/2 (−α − 1/2)(n + 1/2) γ m1 = , n (−α)(n)

(3) (4)

for α < −1/4 and α < −1/2, respectively. Then, using (3) and (4),  α can be determined as the solution of g( α ) − ζ = 0,

(5)

where g( α) = ζ=

 2 (− α − 14 ) (− α )(− α − 12 )

and

m21/2 (n)(n + 12 ) , m1  2 (n + 14 )

and then substituting the value of  α in (3) or in (4) the value of  γ is found. It can be noticed that g( α ) converges asymptotically to one as  α → −∞. As ζ is a random variable that can take values greater than one, there are cases for which (5) does not have a solution. The lower the value of the α parameter, the higher the probability that a solution for (5) does not exist. This issue was solved by Mejail et al. (2003). They needed an estimate of (α, γ ) in every pixel, so

18

Stat Comput (2008) 18: 15–26

Table 1 Statistical description of areas Urban

Forest

Pasture

 α = −1.12,  γ = 229 631.85

 α = −4.74,  γ = 497 304.47

 α = −12.18,  γ = 470 939.78

 1/2 (1, λ)

 α = 0.70, λ = 1.07693e–6  λ = 8.21640e–007

 α = 4.22, λ = 3.19504e–5  λ = 7.00652e–006

 α = 11.59, λ = 2.75324e–4  λ = 1.80084e–005

Size

20 272

5168

3266

0 (α, γ , 1) GA

KA (α, λ, 1)

Fig. 2 Histogram of a pasture region and fitted densities for the GA0 0 (dash-dot line) distributions (solid line),  1/2 (dashed line) and KA

Fig. 3 Histogram of a forest region and fitted densities for the GA0 0 (dash-dot line) distributions (solid line),  1/2 (dashed line) and KA

they used the observations contained in a small box around each coordinate; if  α could not be obtained, the median of surrounding estimates was employed. This is the approach used in this paper whenever local estimates are required. The estimated parameters for urban, forest and pasture regions, along with the sample sizes, are presented in Table 1. It is noticeable that the absolute value of α, the roughness parameter, increases with the homogeneity of the target. The observed image histograms and fittings using the GA0 , KA and  1/2 distributions, are presented in Figs. 2, 3 and 4. It is noticeable that the GA0 distribution is the one that consistently provides the best fit.

3 B-spline representation

Fig. 4 Histogram of an urban region and fitted densities for the GA0 0 (dash-dot line) distributions (solid line),  1/2 (dashed line) and KA

We use the B-spline curve representation for describing object contours in a scene. B-splines are a convenient polynomial representation of curves with the following interesting features:

3. The B-spline approach allows for the local control of the curve by tuning the control points individually. 4. The curve lies within the convex hull induced by the control points.

1. The curve is represented by a few parameters, the control points; this reduces the computational effort to compute it. 2. The order of the polynomial segments is chosen arbitrarily, and it relates to the desired smoothness.

In the following, a brief review of B-spline representation of contours is presented; for more details see Blake and Isard (1998) and Rogers and Adams (1990). Let {Q0 , . . . , QNB −1 } be a set of control points, where Qn = (xn , yn )t ∈ R2 , 0 ≤ n ≤ NB − 1, and let {s0 < s1 <

Stat Comput (2008) 18: 15–26

19

s2 < · · · < sL−1 } ⊂ R be a set of L knots. A B-spline curve of order d is defined as a weighted sum of NB polynomial basis functions Bn,d (s) of degree d − 1, within each interval [si , si+1 ] with 0 ≤ i ≤ L − 1. The spline function is r(s) = (x(s), y(s))t , 0 ≤ s ≤ L − 1, being r(s) =

N B −1

Bn,d (s)Qn ,

n=0

and x(s) = Bt (s)Qx ,

(6)

y(s) = Bt (s)Qy

(7)

where the basis functions vector B(s) of NB components is given by B(s) = (B0,d (s), . . . , BNB −1,d (s))t . The weight vectors Qx and Qy give the first and second components of Qn , respectively. The curves used in this work are closed, with d = 3 or d = 4, and are specified by periodic B-spline basis functions. We now present a brief review of the problem of determining a polygon that generates a fitting B-spline curve with known number of control points, NB . Consider {D0 , D1 , . . . , Dk−1 } ∈ R2 , k points in the image plane, where Di = (xi , yi )t , 0 ≤ i ≤ k − 1. The spline curve that best fits them is sought. Equations (6) and (7), imply that the components Di must satisfy xi = Bt (ti )Qx and yi = Bt (ti )Qy for certain values of ti , 0 ≤ i ≤ k − 1 and NB ≤ k. This linear system is more compactly written in matrix form as D = K(Qx Qy ), where the k × NB elements of the real matrix K are given by Kij = Bj,d (ti ), 0 ≤ i ≤ k − 1, 0 ≤ j ≤ NB − 1, and D = (D0 , D1 , . . . , Dk )t . In the most general case NB < k and, therefore, K is not a square matrix. In this case, the pseudo-inverse matrix form (Qx Qy ) = K + D is used to find the B-spline fitting curve. A useful set of values for the parameters {t0 , . . . , tk−1 } is given by  i=1 Di − Di−1 t0 = 0, t = k−1 ,  ≥ 1. i=1 Di − Di−1 The knot set to build the B-spline basis functions is arbitrarily chosen. The decision as to how many control points are adequate for general region fitting has been taken on an ad hoc basis: we used twenty points in all our experiences with good results. Less points produce coarse regions, while many more than that slow the procedure without noticeable benefits. A more profound study, that could be based, for example, on the minimum description length (MDL) principle (see Figueiredo et al. 2000), would be a desirable automatic adaptive approach in real applications.

4 Boundary detection In this section we present the methods developed to detect region boundaries in speckled imagery. Let E be a scene made up by a background B and a set of N reg distinct regions {R1 , R2 , . . . , RN reg } with boundaries {∂R1 , . . . , ∂RN reg }, respectively. Each of these regions and the background are considered to be random fields of independent, identically distributed, random variables obeying the GA0 distribution and characterized by the values of their parameters (αh , γh ), 1 ≤ h ≤ Nreg. The only assumption we make is that regions and background have different textures, i.e., if α0 is the background roughness then αh = α0 for every 1 ≤ h ≤ N reg. For each region Rh , we want to find the curve Ch that fits the boundary ∂Rh in the image. As shown in Sect. 2.1, there is a correspondence between the various areas present in the image and the parameters of roughness and scale. From this correspondence a classification of the image can be done. The algorithms we propose aim at separating regions of a certain specified type from the rest of the image. We first obtain a rough approximation, or seed: the starting regions of interest. The contour extraction algorithms work over these regions instead of analyzing the whole image reducing, thus, the computational effort. This initial region detection is computed over nonoverlapping blocks of the image using as additional input the number of blocks and the type of region to be detected (homogeneous, heterogeneous or extremely heterogeneous); the details are presented in Gambini et al. (2006). This step is crucial, and prior information about the typical size of objects to be detected should be provided. Once Nreg regions are identified, each centroid ch , 1 ≤ h ≤ Nreg, is computed. If a point belongs to the object boundary, then a sample taken from its neighborhood should exhibit a change in its properties and, therefore, it could be considered as a transition point. In order to find transition points, N segments s (i) , 1 ≤ i ≤ N , of the form s (i) = c pi are considered for each region, c being the centroid of the initial region, and pi an extreme point outside of the region. It is necessary for the centroid c to be in the interior of the object whose contour is (i) sought. A strip Sh is defined over each segment s (i) . This procedure is illustrated in Fig. 5: to the left with a fluxogram, and to the right with data. In order to find the transition points we use five methods: three that employ the raw data (Maximum Discontinuity, Fractal Dimension and Maximum Likelihood) and two that use the estimated roughness (Maximum Discontinuity and Anisotropic Diffusion). The width of the strips Sh(i) was determined analyzing real data and seeking for the smallest sample size leading to few situations with numerical instabilities. Squared windows of side 20 proved being a good compromise, so strips

20

Stat Comput (2008) 18: 15–26

Fig. 5 Sequence of operations leading from the data conditioned on the blocks (Z | B) to the N strips Sh(i) , 0 ≤ i ≤ N − 1, for each region 1 ≤ h ≤ Nreg

of this width were employed in all the following situations. When no estimation was possible, the afore mentioned technique by Mejail et al. (2003) was used. Frery et al. (2004) show that specialized algorithms are needed to extract information from speckled data using small samples. 4.1 Maximum discontinuity on raw data The general problem consists of finding a transition point (i) within each strip Sh . These points will be sought using the data along the discrete version of s (i) , for which we will not (i) ), introduce a new notation (see Fig. 5): s (i) = (z1(i) , . . . , zm 1 ≤ i ≤ N. A way of finding a transition point is seeking for a change in s (i) ; in order to do so, following (Blake and Isard 1998), we apply a high pass filter and identify the maximum value.

Thus, the method consists of finding the position of maximum change in the filtered data, which is shown in Algorithm 1. Algorithm 1: Edge detection by maximum discontinuity using raw data 1. for each line s (i) (a) Convolve the data with the mask [−2, −1, 0, 1, 2] and find the position that corresponds to the maximum discontinuity among the values in the array s (i) . (b) The point found bi corresponds to the transition point between regions. 2. endfor 3. Build the B-spline curve that interpolates the boundary points {b1 , . . . , bN }.

Stat Comput (2008) 18: 15–26

21

A high-pass filter enhances local variation, hampering even further the edge detection in the presence of noise. It comes as no surprise that this method provides the worst results, though at a very low computational cost. 4.2 Fractal dimension The fractal dimension characterizes the geometric complexity and the small-scale variations of a set. The reader is referred to the works by Falconer (1990) and Peitgen and Saupe (1986), among others, for further details. Among the available methods for computing the fractal dimension, Falconer (1990) proposes the box-counting technique. It is reported as a good compromise between precision and computational cost, and it requires the input image to be binary. Let U ⊂ Rn be a non-empty set and define |U | = sup{|x − y| : x, y ∈ U } its diameter. Denote {Ui }i∈I a countable class of sets in Rn such that for every i ∈ I holds that  sup{|Ui |} ≤ r, then for every F ⊂ i∈I Ui , {Ui }i∈I is a rcover for F . Consider the closed set A ⊂ Rn and let Nr (A) be the smallest number of sets with diameter at most r that cover A, then the box-counting dimension of A is given by BC = lim

r→0

log(Nr (A)) log( 1r )

,

(8)

if the limits exists. If A is a planar set, this dimension is equivalent to using squared cells of side r; the box-counting dimension of such set is related to the number of cells where the set has non-null measure. Let r be the side of the cell used to cover the set, and Nr the number of cells where the set has non-null measure, then, if the limit exists, the box-counting dimension is BC = lim

r→0

log(Nr ) log( 1r )

.

This measure can be easily computed since it uses squared boxes, whereas the one given by equation (8) employs arbitrarily shaped sets. The procedure for computing the BC of a binary image consists of partitioning the image in squared cells of side r and counting Nr , the number of cells that cover the object. This is performed for several values of r, and BC can be estimated as the slope of the straight line that best fits the points (log(Nr ), log( 1r )). Classical regression is usually employed to compute this estimate. Since the images under analysis are real-valued, they have to be transformed into binary data before computing BC. The ideal threshold for this operation is the value that separates objects from background, and the most successful techniques are adaptive (Wang and Bai 2003). In order to keep the computational cost low, a global threshold is used.

As can be seen in Fig. 1, the higher the value of γ the brighter the image, so we estimated this parameter in the data under analysis and used a threshold empirically tuned by this estimate in order to obtain a binary image that retains enough information about the objects. We estimate γ from the whole dataset and then use 3/(2 γ 1/2 ) as threshold. We noted that windows of side 6 × 6 provide enough information for estimating BC while reducing the computational cost of using more data. The method, adapted from Chen et al. (1993) for our problem, consists of the steps described in Algorithm 2. Algorithm 2: Edge detection by box-counting dimension estimation 1. for each line s (i) (a) Binarize the data in the strip S (i) using a threshold that depends on the estimated scale parameter γ . (b) Consider a window of size 6 × 6 at each point of s (i) and estimate the box-counting dimension. Store the (i)  (i) , . . . , BC  m  (i) = [BC ]. estimates in an array BC 1 (i) (c) Find bi , the position on line s that corresponds to the maximum discontinuity among the values in  (i) convolving with the mask [−2, −1, the array BC 0, 1, 2]. 2. endfor 3. Build the B-spline curve that interpolates the boundary points {b1 , . . . , bN }. 4.3 Maximum likelihood For each segment s (i) , 1 ≤ i ≤ N , we consider the following partition (i)

Zk ∼ GA0 (αr , γr ),

k = 1, . . . , j,

(i) Zk

k = j + 1, . . . , m

∼ GA0 (αb , γb ),

where for each k, with 1 ≤ k ≤ m, zk(i) is the realization (i) of the random variable Zk . The parameters (αr , γr ) and (αb , γb ) characterize the region and its background, respectively, and are unknown. Every estimation is performed using local information. The data employed are the pixels within a rectangular window with its major axis coinciding with the segment under analysis, and such that it includes samples of both foreground and background. The rectangles inside the window depict the areas where data are collected in order to estimate the quantities that lead the edge detection procedure. The transition point of each segment s (i) is the point maximizing the likelihood L(j ) =

j  i=1

Pr(zi ; αr , γr )

m  i=j +1

Pr(zi ; αb , γb ).

22

Stat Comput (2008) 18: 15–26

Alternatively, we can maximize (j ) = ln L(j ) =

j 

is obtained. If a point lies on the boundary between two regions, then it exhibits an abrupt change in the values of the α estimates. Thus, the method consists of finding the position for maximum change in  α , which is shown in Algorithm 4.

ln(fG 0 (zi ; αr , γr ))

i=1

+

m  i=j +1

Algorithm 4: Edge detection by maximum discontinuity using roughness estimates

ln(fG 0 (zi ; αb , γb )).

According to (1) (j ) =

j 

ln

2nn  (n − αr ) zi2n−1



γrαr  (−αr )  (n) (γr + nzi2 )n−αr

m  2nn  (n − αb ) zi2n−1 + ln . γbαb  (−αb )  (n) (γb + nzi2 )n−αb i=j +1

i=1

Finally, the estimated index on the segment that corresponds to the transition point  j is given by  j = arg max (j ). j

(9)

The scheme of this procedure is shown in Algorithm 3. Algorithm 3: Edge detection by maximum likelihood using raw data 1. for each segment s (i) , i = 1, . . . , N (a) Estimate the parameters (αr , γr ) and (αb , γb ). (b) Find the index  j on the segment s (i) that maximizes equation (9); it corresponds to the border point bi in the image. 2. endfor 3. Build the B-spline curve that interpolates the boundary points {b1 , . . . , bN }. An evaluation of this method using a family of different synthetic star-shaped regions was carried out in Gambini et al. (2006). It exhibits a very good performance from the point of view of the mean square error. Other methods for the maximization of a Likelihood function have been proposed by various authors in the context of image analysis (see, for instance, Jain 1989; Lim 1989) but the technique proposed in this work proved being the best suited for the situation at hand. 4.4 Maximum discontinuity on roughness estimates Another way of finding the transition point on a line s (i) is to estimate the α parameter in a window centered on each pixel on the line using the data on the strip S (i) . Then an (i) (i) (i) = [ α1 , . . . , αm ] of estimates of the α parameter array

1. for each line s (i) (a) Estimate the parameter α for each pixel on s (i) us(i) = ing a sliding window. This generates an array

(i) (i) [ α1 , . . . , αm ] of estimated values of α. (b) Find bi , the position on line s (i) that corresponds to the maximum discontinuity among the values in (i) convolving with the mask [−2, −1, the array

0, 1, 2]. 2. endfor 3. Build the B-spline curve that interpolates the boundary points {b1 , . . . , bN }. With this method, a small window causes the estimation to be too noisy and numerically unstable (a detailed account of this issue is discussed by Frery et al. 2004) and a large window increases the computational cost and diminishes the discriminatory power (resolution of the procedure). In order to have acceptable estimation at moderate computational cost we used 9 × 9 windows and the estimation technique presented in Sect. 2.2. In order to alleviate the spurious variations that may lead to a wrong decision, next algorithm incorporates a smoothing step that was designed to preserve small details. 4.5 Anisotropic diffusion The modeling of speckled data suggest that the core information any boundary detection needs resides in the statistical parameters. These estimates convey the information, but it may be masked by random fluctuation, so smoothing is desirable provided it does not introduce distortions. Perona and Malik (1990) proposed an algorithm that combats noise preserving boundary features. In its continuous version, it consists of producing a sequence of images I (·, ·, t), t ≥ 0, according to ∂I (x, y, t)/∂t = ∇ · [g( ∇I )∇I ], where I (·, ·, 0) : R2 → R+ is the original image, t is an artificial time parameter, ∇I is the image gradient, ∇I is the image gradient magnitude, ‘∇ · ’ denotes the divergence, g : R → [0, 1] is an edge detection function with the only constraints that (i) g(x) → 0 monotonically when x → ∞, and (ii) g(x) → 1 when x → 0. More details can be found in, among others, Weickert (1998). The position of the discontinuity on s (i) is found by convolving the smoothed roughness estimates with a cyclic

Stat Comput (2008) 18: 15–26

border detection operator, followed by a convenient thresholding. The scheme of this procedure is shown in Algorithm 5. Algorithm 5: Edge detection by anisotropic diffusion 1. for each s (i) (a) Estimate the parameter α for each pixel on s (i) us(i) = ing a sliding window. This generates an array

(i) (i) [ α1 , . . . , αm ] of estimated values of α. (i) using anisotropic diffusion. (b) Smooth the array

(i) . This generates the smoothed estimates array

S (i) (c) Find bi , the position on line s that corresponds to the maximum discontinuity among the values in (i) convolving with the mask the smoothed array

S [−2, −1, 0, 1, 2]. 2. endfor 3. Build the B-spline that interpolates the points {b1 , . . . , bN }. Once five edge detection techniques have been proposed, we proceed to their comparative assessment.

5 Methodology for error evaluation In this section we present the methodology employed for comparing the five edge detection techniques previously discussed. Two important issues have to be defined: the error measure and the general setup for the assessment. Since we are dealing with B-splines, which are determined by control points, a convenient measure of error can be built comparing the position of estimated and true control points; this requires the definition of true regions. Using a single image for this assessment would lead to meaningless results. Since we are interested in a general comparison, and since we have a sound statistical model for the data, a Monte Carlo experiment will be devised to compute approximate mean errors in a variety of situations of interest. In order to compute the error in estimating by b the true boundary point PT , for each of the five aforementioned methods, a phantom image was used. This image is a 20 × 100 pixels data set divided in halves: a background and a region. Each half is filled with samples from a different GA0 law. Three situations were simulated. Each situation is an image with halves where samples from the GA0 (α, 1, 1) distribution are observed: (α1 , α2 ) = (−1.1, −3), (α1 , α2 ) = (−1.1, −10) and (α1 , α2 ) = (−3, −10), respectively. These values reflect the noisiest possible image (one look) and roughness corresponding to different types of areas, namely

23

extremely heterogeneous, heterogeneous and homogeneous, as reported by Gambini et al. (2006). Two hundred replications were made for each situation, the five algorithms were applied to each of these images, the transition points were estimated and the error was computed for each algorithm in each replication. The distance of these points to the true boundary was evaluated, from which the empirical distribution of the error was computed for each situation. Denote by H (k) the number of replications for which the error is smaller than k pixels, then an estimate of this probability is given by f (k) = H (k)/200, for k ≥ 0. If for any method M it is true that f M (2) < 0.5, then this method must be discarded, because in this case less than half the number of transition points have been found with an accuracy of 2 pixels, which is unacceptable. A method M is more efficient than other method K if f M (k) > f K (k), for values of k that are close to zero. Next section relates the results for 0 ≤ k ≤ 4.

6 Results In this section, we will show that boundary detection using maximum likelihood for the GA0 distribution (Algorithm 3) surpasses all other algorithms in the situations here considered. These situations depict typical cases of boundaries between, e.g., urban areas, forests and pastures. Figure 6 shows, for each of the three situations, f M (k): the estimated probability that method M (maximum discontinuity using raw data, fractal dimension, maximum likelihood, maximum discontinuity on the roughness and anisotropic diffusion) finds the correct edge with an error of less that 0 ≤ k ≤ 4 pixels, being f M (0) the probability of a correct detection. The situations here considered are edges between (a) extremely heterogeneous and heterogeneous, (b) extremely heterogeneous and homogeneous, and (c) heterogeneous and homogeneous data. For instance, Fig. 6(c) shows that f AD (1) = 0.660 and that f ML (1) = 0.945; this means that the probability that the maximum likelihood method estimates an edge between forest and pasture with an error of at most two pixels is about 1.5 times bigger than the chance anisotropic diffusion has. Note that the estimated probability that maximum likelihood finds the correct edge with an error of less that 0 ≤ k ≤ 4 pixels is always close to 1. The only situations for which it is not 1 are for k = 0, 1, with (a) (α1 , α2 ) = (−1.1, −3), for which it is 0.95, 0.99, and (c) (α1 , α2 ) = (−3, −10), for which it is 0.995. The results obtained by Monte Carlo numerical experiments show that the maximum likelihood edge detection is consistently better than the other four methods, being the

24

Stat Comput (2008) 18: 15–26

Fig. 6 Relative frequency curves f M (k) in the three situations for the five techniques

detection of discontinuities on raw data the worst (the corresponding curves are always at the bottom, so the probability of correctly finding an edge is almost zero). Anisotropic diffusion is the second best in every case but for k = 0 in the first situation. The other three techniques are not competitive. It is noteworthy that the two best methods employ different information: the former uses raw data to compute a likelihood function, and the latter uses a smoothed version of the estimated roughness. In order to obtain good results, the parameters used for anisotropic diffusion were σ = 100, λ = 0.25, t = 10; these parameters impose a high computational cost on Algorithm 5. This last algorithm is at least ten times and at worst one hundred times slower than maximum likelihood, so Algorithm 3 produces the best edge detection requiring one or two orders of magnitude less of the computational time demanded by the next best procedure. A 256 × 256 synthetic image obeying the GA0 (α, 1, 1) model, with three different city-like areas with α = −1.5, −2.1, −2.7 and a background with α = −10 was used to assess the robustness of Algorithm 3 with respect to initialization. All of the parameters are considered unknown and they are estimated from the image. Initial boundaries (thin lines) and final boundaries (thick lines) can be seen in Fig. 7. Three types of initialization were used: across region boundaries (upper left), inside region (upper right) and outside region (lower middle). In every case, Algorithm 3 performs well regardless the seed. In the following, the results of applying the maximum likelihood technique to three real images is shown. In every

Fig. 7 Results of Algorithm 3 with three types of starting regions

case, the images are displayed with contrast enhancement since the original data reveal little boundary information. The detection is applied on the original data, though. Figure 8 shows the result of applying maximum likelihood edge detection to a Radarsat image over flooded fields, namely an area over the Paraná river delta, Argentina: the thin line is the seed and the think one the detected edge. In this case, large homogeneous areas were specified as target, and the algorithm converged to a sensible result regardless the noise and the artifacts that are visible inside the detected region. Figure 9 shows the detection of an edge between secondary (darker) and primary forest (lighter) in Tapajós, Brazilian Amazon forest. This detection is quite difficult using only imagery in the visible spectrum, since the canopies look alike. Though the contrast of a SAR image reveals this information, it is the different textures that provide enough

Stat Comput (2008) 18: 15–26

25

Fig. 10 Detected edge in an ESAR image: grass and trees Fig. 8 Detected edge in a Radarsat image: flooded area

Fig. 9 Detected edge in a Radarsat image: secondary and primary forest

information in order to establish an edge. The outer line is the seed, while the inner is the detected contour. The equivalent number of looks of both RADARSAT imagens was 3.04. Figure 10 shows an ESAR image, with equivalent number of looks of 3.12, and the result of applying the algorithm to the detection of an area of grass surrounded by trees in a suburban region of München, Germany. The initial region is the small polygon inside the final edge. In order to quantify the difficulty of segmenting real areas, a general test statistic for contrasting the null hypothesis of equality of means against the alternative of different means can be used. If zR and zB denote samples from the region and the surrounding background, a measure of separability between them is tR,B =

( zR − zB )2 , sz2R + sz2B

(10)

where z and sz denote, respectively, the sample mean and standard deviation of sample z. The smaller this measure, the harder the task of finding a boundary between the region and its background. As presented by Gambini et al. (2006), the separability indexes in real images range from 1.18 to 4.19. In that paper, the authors simulated SAR imagery with a variety of shapes and parameters; using (2) in the expression presented in (10) one can compute the theoretical separability between background and foreground. The easiest simulated case corresponds to n = 5, αR = −3, γR = 1, αB = −10 and γB = 1, yielding tR,B = 1.21; this level of complexity is comparable with the hardest real data case. The hardest simulated case is when n = 1, αR = −7, γR = 1, αB = −10 and γR = 1, for which tR,B = 0.02. In this way, our simulation study presented results related to problems that are at least as hard as the ones found in practice.

7 Conclusions and future work Five approaches of active contour region fitting in speckled imagery were compared. These algorithms use strips of information, either raw data (maximum discontinuity, fractal dimension and maximum likelihood) or estimated roughness (maximum discontinuity and anisotropic diffusion) to find edges in speckled data. The error in finding edges was defined, and a Monte Carlo experiment was used to assess this error in a variety of situations that appear in the practice of speckled imagery analysis. Edge detection by maximum likelihood using raw data proved being consistently superior than the other four techniques, with the best computational cost with respect to its next competitor, edge detection on roughness estimates smoothed by anisotropic diffusion.

26

The robustness of the best approach with respect to the starting edge was also assessed, and its overall behavior makes maximum likelihood on raw data the best candidate for edge detection on speckled data. The use of robust (see Allende et al. 2006; Bustos et al. 2002; Lucini et al. 2003) and improved (see Cribari-Neto et al. 2002; Vasconcellos et al. 2005) estimators is under assessment. The best performing technique can be easily adapted to any kind of image, provided a sound and tractable statistical model for the observed data. Acknowledgements The authors are grateful to CNPq, FAPEAL and SeCyT for partial funding. Haydee Karszenbaum, Institute for Astronomy and Space Physics (IAFE), Argentina, kindly provided the image over the Paraná river.

References Allende, H., Frery, A.C., Galbiati, J., Pizarro, L.: M-estimators with asymmetric influence functions: the GA0 distribution case. J. Stat. Comput. Simul. 76(11), 941–956 (2006) Barndorff-Nielsen, O.E., Blæsild, P.: Hyperbolic distributions. In: Kotz, S., Johnson, N.L. (eds.) Encyclopedia of Statistical Sciences, vol. 3. pp. 700–707. Wiley, New York (1983) Blackledge, J., Fowler, E.: Fractal dimensions segmentation of Synthetic Aperture Radar images. In: International Conference on Image Processing and its Applications, IEEE, pp. 445–449 (1992) Blake, A., Isard, M.: Active Contours. Springer, Berlin (1998) Brigger, P., Hoeg, J., Unser, M.: B-spline snakes: A flexible tool for parametric contour detection. IEEE Trans. Image Process. 9(9), 1484–1496 (2000) Bustos, O.H., Lucini, M.M., Frery, A.C.: M-estimators of roughness and scale for GA0-modelled SAR imagery. EURASIP J. Appl. Signal Process. 2002(1), 105–114 (2002) Chen, S.S., Keller, J.M., Crownover, R.M.: On the calculation of fractal features from images. IEEE Pattern Anal. Mach. Intell. 15(10), 1087–1090 (1993) Cipolla, R., Blake, A.: The dynamic analysis of apparent contours. In: Proceedings of the 3rd Int. Conf. on Computer Vision, pp. 616– 625 (1990) Cribari-Neto, F., Frery, A.C., Silva, M.F.: Improved estimation of clutter properties in speckled imagery. Comput. Stat. Data Anal. 40(4), 801–824 (2002) Dennis, T.J., Dessipris, N.G.: Fractal modelling in image texture and analysis. IEE Proc. F 136(5), 227–235 (1989) Du, G., Yeo, T.S.: A novel multifractal estimation method and its application to remote image segmentation. IEEE Trans. Geosci. Remote Sens. 40, 980–982 (2002) Falconer, K.: Fractal Geometry: Mathematical Foundations and Applications. Wiley, Chichester (1990) Figueiredo, M., Leitão, J.: Bayesian estimation of ventricular contours in angiographic images. IEEE Trans. Med. Imaging 11(3), 416– 429 (1992) Figueiredo, M., Leitão, J., Jain, A.K.: Unsupervised contour representation and estimation using B-splines and minimun description lenght criterion. IEEE Trans. Image Process. 9(6), 1075–1087 (2000) Frery, A.C., Müller, H.-J., Yanasse, C.C.F., Sant’Anna, S.J.S.: A model for extremely heterogeneous clutter. IEEE Trans. Geosci. Remote Sens. 35(3), 648–659 (1997) Frery, A.C., Cribari-Neto, F., Souza, M.O.: Analysis of minute features in speckled imagery with maximum likelihood estimation. EURASIP J. Appl. Signal Process. 2004, 2476–2491 (2004)

Stat Comput (2008) 18: 15–26 Gambini, J., Mejail, M.E., Jacobo-Berlles, J., Frery, A.C.: Feature extraction in speckled imagery using dynamic B-spline deformable contours under the G0 model. Int. J. Remote Sens. 27(22), 5037– 5059 (2006) Germain, O., Réfrégier, P.: Edge location in SAR images: performance of the likelihood ratio filter and accuracy improvement with an active contour approach. IEEE Trans. Image Process. 10(1), 72– 78 (2001a) Germain, O., Réfrégier, P.: Statistical active grid for segmentation refinement. Pattern Recognit. Lett. 22(10), 1125–1132 (2001b) Goodman, J.W.: Some fundamental properties of speckle. J. Opt. Soc. Am. 66, 1145–1150 (1976) Jain, A.K.: Fundamentals of Digital Image Processing. Prentice-Hall, Englewood Cliffs (1989) Jørgensen, B.: Statistical Properties of the Generalized Inverse Gaussian Distribution. Lecture Notes in Statistics, vol. 9. Springer, New York (1982) Keller, T.: Texture description and segmentation through fractal geometry. Comput. Vis. Graph. Image Process. 45, 150–166 (1989) Lim, J.S.: Two-dimensional Signal and Image Processing. Prentice Hall Signal Processing Series. Prentice Hall, Englewood Cliffs (1989) Liu, Y., Li, Y.: Image feature extraction and segmentation using fractal dimension. In: International Conference on Information and Signal Processing, IEEE, pp. 975–979 (1997) Lucini, M.M., Ruiz, V.F., Frery, A.C., Bustos, O.H.: Robust classification of SAR imagery. In: IEEE International Conference on Acoustics, Speech and Signal Processing, IEEE, Hong Kong, pp. 557–560 (2003) Manski, C.F.: Analog Estimation Methods in Econometrics. Monographs on Statistics and Applied Probability, vol. 39. Chapman & Hall, New York (1988), url: http://elsa.berkeley.edu/books/ analog.html Medioni, G., Yasumoto, Y.: Corner detection and curve representation using curve B-splines. In: Proc. Conf. Computer Vision and Pattern Recognition, pp. 764–769 (1986) Mejail, M.E., Frery, A.C., Jacobo-Berlles, J., Bustos, O.H.: Approximation of distributions for SAR images: Proposal. evaluation and practical consequences, Lat. Am. Appl. Res. 31, 83–92 (2001) Mejail, M., Jacobo-Berlles, J., Frery, A.C., Bustos, O.H.: Classification of SAR images using a general and tractable multiplicative model. Int. J. Remote Sens. 24(18), 3565–3582 (2003) Oliver, C.J.: Information from SAR images. J. Phys. D: Appl. Phys. 24, 1493–1514 (1991) Peitgen, H.O., Saupe, D.: The Science of Fractal Images. Springer, Berlin (1986) Peli, T.: Multiscale fractal theory and object characterization. J. Opt. Soc. Am. 7(6), 1113–1123 (1990) Perona, P., Malik, J.: Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell. 12(7), 629–639 (1990) Quartulli, M., Datcu, M.: Stochastic geometrical modelling for built-up area understanding from a single SAR intensity image with meter resolution. IEEE Trans. Geosci. Remote Sens. 42(9), 1996–2003 (2004) Rogers, D.F., Adams, J.A.: Mathematical Elements for Computer Graphics, 2 edn. McGraw-Hill, New York (1990) Seshadri, V.: The Inverse Gaussian Distribution: A Case Study in Exponential Families. Claredon Press, Oxford (1993) Vasconcellos, K.L.P., Frery, A.C., Silva, L.B.: Improving estimation in speckled imagery. Comput. Stat. 20(3), 503–519 (2005) Wang, L.S., Bai, J.: Threshold selection by clustering gray levels of boundary. Pattern Recognit. Lett. 24(12), 1983–1999 (2003) Weickert, J.: Anisotropic Diffusion in Image Processing. Teubner, Stuttgart (1998)

Accuracy of edge detection methods with local ... - Springer Link

Sep 11, 2007 - which regions with different degrees of roughness can be characterized ..... Among the available methods for computing the fractal dimension ...

786KB Sizes 1 Downloads 296 Views

Recommend Documents

Localization and registration accuracy in image guided ... - Springer Link
... 9 January 2008 / Accepted: 23 September 2008 / Published online: 28 October 2008. © CARS 2008. Abstract. Purpose To measure and compare the clinical localization ..... operating room, the intraoperative data was recorded with the.

An Approach for the Local Exploration of Discrete ... - Springer Link
Optimization Problems. Oliver Cuate1(B), Bilel Derbel2,3, Arnaud Liefooghe2,3, El-Ghazali Talbi2,3, and Oliver Schütze1. 1. Computer Science Department ...

Local Stability of an Endoreversible Heat Engine ... - Springer Link
We present a local stability analysis of an endoreversible engine working in an ..... f. (b). (a). Max. Power. Max. Ecological. Max. Power. Max. Ecological. Fig.

Temporal Representation in Spike Detection of Sparse ... - Springer Link
and stream mining within a framework (Section 2.1): representation of time with ... application data sets (Section 3.2), experimental design and evaluation ...

Foveal Algorithm for the Detection of Microcalcification ... - Springer Link
Ewert House, Ewert Place, Summertown, Oxford OX2 7BZ, UK ... present in about a quarter of the total number of screening mammograms. For these reasons ... microcalcifications are expected to appear bright in a mammogram, indeed to be.

Context Driven Focus of Attention for Object Detection - Springer Link
an image. In computer vision, object detectors typically ignore this in- formation. ... detection in urban scenes using a demanding image database. Results.

Spatial methods for plot-based sampling of wildlife ... - Springer Link
Sep 19, 2007 - The data are assumed to come from a spatial stochastic ..... Patterson HD, Thompson R (1971) Recovery of interblock information when block ...

Spatial methods for plot-based sampling of wildlife ... - Springer Link
Sep 19, 2007 - Monitoring ecological populations is an important goal for both ... units, call it τ(z), where z is a vector of the realized values of a spatial ..... The distance between sample units was computed in kilometers from the center.

LNCS 6622 - NILS: A Neutrality-Based Iterated Local ... - Springer Link
a new configuration that yields the best possible fitness value. Given that the .... The neutral degree of a given solution is the number of neutral solutions in its ...

LNCS 3973 - Local Volatility Function Approximation ... - Springer Link
S&P 500 call option market data to illustrate a local volatility surface estimated ... One practical solution for the volatility smile is the constant implied volatility approach .... Eq. (4) into Eq. (2) makes us to rewrite ˆσRBF (w; K, T) as ˆσ

XML schema refinement through redundancy detection ... - Springer Link
Feb 20, 2007 - egy to avoid data redundancies is to design redundancy-free schema from the ...... nodesCN,C A, andCNA are removed becauseC is an XML. Key. ..... the name element of province and the country attribute of city together ...

Computer-Aided Polyp Detection for Laxative-Free CT ... - Springer Link
propose an automated algorithm to subtract stool prior to the computer aided detection of colonic ... [4], material classes were detected via hard thresholds.

LNCS 4233 - Fast Learning for Statistical Face Detection - Springer Link
Department of Computer Science and Engineering, Shanghai Jiao Tong University,. 1954 Hua Shan Road, Shanghai ... SNoW (sparse network of winnows) face detection system by Yang et al. [20] is a sparse network of linear ..... International Journal of C

A Bayesian approach to object detection using ... - Springer Link
using receiver operating characteristic (ROC) analysis on several representative ... PCA Ж Bayesian approach Ж Non-Gaussian models Ж. M-estimators Ж ...

Interiors of Sets of Vector Fields with Shadowing ... - Springer Link
Corresponding to Certain Classes of Reparameterizations. S. B. Tikhomirov. Received May 18, 2008. Abstract—The structure of the C1-interiors of sets of vector ...

Data integration with uncertainty - Springer Link
Nov 14, 2008 - sources by automatic methods (e.g., HTML pages, emails, blogs). ..... If a tuple is an answer to Q under multiple mappings in m, then we add.

Isoperimetric inequalities for submanifolds with ... - Springer Link
Jul 23, 2011 - if ωn is the volume of a unit ball in Rn, then. nnωnVol(D)n−1 ≤ Vol(∂D)n and equality holds if and only if D is a ball. As an extension of the above classical isoperimetric inequality, it is conjectured that any n-dimensional c

Calculus of Variations - Springer Link
Jun 27, 2012 - the associated energy functional, allowing a variational treatment of the .... groups of the type U(n1) × ··· × U(nl) × {1} for various splittings of the dimension ...... u, using the Green theorem, the subelliptic Hardy inequali

Plant location with minimum inventory - Springer Link
fractional solution we were able to derive integer solutions within 4% of optimality. ... F. Barahona, D. Jensen / Mathematical Programming 83 (1998) 101-111.

A link between complete models with stochastic ... - Springer Link
classical ARCH models, a stationary solution with infinite variance may exists. In ..... must compute the required conditional expectations and variances. Setting ...

Combining Metaheuristics and Exact Methods for ... - Springer Link
and a call to the exploration program. This program ... granularity management or termination detection, exchanging lists of nodes on a computational grid is ...

Augmented Lagrangian Method, Dual Methods and ... - Springer Link
Abstract. In the recent decades the ROF model (total variation (TV) minimization) has made great successes in image restoration due to its good edge-preserving property. However, the non-differentiability of the minimization problem brings computatio

Practical methods for constructing suffix trees - Springer Link
Sep 26, 2005 - Richard A. Hankins · Jignesh M. Patel. Practical methods for ..... reduces the suffix array construction using a two-thirds–one- thirds split of the ...

New Exact Solution of Dirac-Coulomb Equation with ... - Springer Link
Sep 19, 2007 - brings on that the solutions of the Klein-Gordon equation and the Dirac ... and its magnetic moment in a completely natural way and so on.