Journal of the Physical Society of Japan Vol. 77, No. 5, May, 2008, 054803 #2008 The Physical Society of Japan

Bayesian-Optimal Image Reconstruction for Translational-Symmetric Filters Satohiro T AJIMA1 , Masato I NOUE2 y, and Masato O KADA1;3 z 1

Department of Complexity Science and Engineering, Graduate School of Frontier Sciences, The University of Tokyo, Kashiwa, Chiba 277-8562 2 Department of Electrical Engineering and Bioscience, School of Science and Engineering, Waseda University, Tokyo 169-8555 3 RIKEN Brain Science Institute, Wako, Saitama 351-0198 (Received January 29, 2008; accepted March 18, 2008; published May 12, 2008)

Translational-symmetric filters provide a foundation for various kinds of image processing. When a filtered image containing noise is observed, the original one can be reconstructed by Bayesian inference. Furthermore, hyperparameters such as the smoothness of the image and the noise level in the communication channel through which the image observed can be estimated from the observed image by setting a criterion of maximizing marginalized likelihood. In this article we apply a diagonalization technique with the Fourier transform to this image reconstruction problem. This diagonalization not only reduces computational costs but also facilitates theoretical analyses of the estimation and reconstruction performances. We take as an example the Mexican-hat shaped neural cell receptive field seen in the early visual systems of animals, and we compare the reconstruction performances obtained under various hyperparameter and filter parameter conditions with each other and with the corresponding performances obtained under no-filter conditions. The results show that the using a Mexican-hat filter can reduce reconstruction error. KEYWORDS: Bayesian inference, image reconstruction, denoising, translational invariance, Fourier transform, hyperparameter estimation, filter design DOI: 10.1143/JPSJ.77.054803

1.

Introduction

2.

Bayesian inference is a useful approach to image restoration problems,1–4) and in this paper we extend it to problems of image reconstruction5) under a translational-symmetric filter. Suppose, for example, that we want to reconstruct an original image that a Mexican-hat shaped filter (an edge detector) has transformed into an edge-enhanced image we observe through a very noisy communication channel and that we know the filter shape but have no exact information about the original image. Given probabilistic models of the original image and communication channel, we can use the Bayes’ theorem to estimate the posterior probability distribution of the original. We can also estimate the model-controlling parameters (hyperparameters), such as the noise level of the communication channel or the smoothness of the original image, by maximizing their marginalized likelihood over the original images. The computational costs of these estimations, however, become a problem when the images are large. In this paper we show that if the filter is linear and translational-symmetric, these computational costs can be reduced by a diagonalization technique utilizing the Fourier transform.6,7) One of our aims in this paper is to provide, for the first time, the diagonalized formulation of Bayesian reconstruction and hyperparameter estimation from filtered images. This diagonalization also enables us to evaluate the theoretical mean value of the reconstruction errors. We demonstrate the hyperparameter estimation and the image reconstruction on the Mexican-hat shaped filter. At the end of the paper, we consider the optimal filter parameter minimizing the reconstruction error. 

E-mail: [email protected] E-mail: [email protected] z E-mail: [email protected] y

Model

2.1 Communication channel and likelihood When  and  are images whose pixel luminance values are arrayed in a vector form and we observe the original image  as the image  after a transformation process, the problem of estimating  from  can be represented as a calculation of the posterior probability following the Bayes’ theorem: PðjÞ ¼

PðjÞPðÞ PðjÞPðÞ ¼Z : PðÞ d PðjÞPðÞ

ð1Þ

The observation model we consider here is a communication channel defined by the noise n and the transformation matrix A: A

 !  ¼ A þ n:

ð2Þ

Note that when A is the identity matrix the problem is a conservative image restoration problem. If the observation noise is additive white Gaussian noise the likelihood of  at a given  is Pðj; A; RÞ ¼ Z

exp½H L ð; ; A; RÞ

;

ð3Þ

d exp½H L ð; ; A; RÞ

where H L ð; ; A; RÞ  R ð  AÞT ð  AÞ;

ð4Þ

and R is a scalar value inversely proportional to the noise variance (the smaller the R, the higher the noise level). T denotes transposition: X T ¼ ½Xr;q  for X ¼ ½Xq;r , where q and r are indices corresponding to pixel loci. We can calculate the integral in the denominator on the right-hand side of eq. (3) as follows:

054803-1

J. Phys. Soc. Jpn., Vol. 77, No. 5

Z

S. TAJIMA et al.

d exp½H L ð; ; A; RÞ ¼

pffiffiffiffiffiffiffiffiffiffiffiffiffiffi N =RN ;

with ð5Þ

Hð; ; A; Þ  H L þ H P  log pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi  log N =jUj

where N is the number of pixels in the image.

¼ ð  V 2.2 Prior distribution of the original image To calculate the posterior probability in eq. (1), we need some a priori knowledge about the original image . Assuming original images to be smooth ones,8,9) we model the prior generative probability distribution of  as follows:7) Pðj; hÞ ¼ Z

exp½H P ð; ; hÞ

; d exp½H P ð; ; hÞ X X X ðq  r Þ2 þ h q 2 : H P ð; ; hÞ   q r2BðqÞ

ð6Þ

ð7Þ

q

Here  and h are positive scalars. A larger  means that smoother images are more likely, and h is a supplemental parameter for constraining the pixel luminance close to zero. For simplicity of the subsequent formulations here, the mean luminance of the original image is assumed to be normalized to zero. If necessary we could instead introduce another parameter for mean luminance. When we define an image as a d-dimensional hypercubic lattice of pixel arrays and BðqÞ as a set of the 2d nearest neighbor pixels of q, the length of each side of which is N 1=d (i.e., N pixels in an image), the first term on the right-hand side of eq. (7) is X X ðq  r Þ2 ¼  T J ; ð8Þ q r2BðqÞ

where  is Kronecker’s delta and  runs over d-dimensional orthonormal bases:  2 ð1; 0; . . . ; 0Þ; ð0; 1; . . . ; 0Þ; . . . ; ð0; . . . ; 0; 1Þ. When we define a positive definite symmetric matrix U  J þ hI, we can write Hð; ; hÞ as H P ð; ; hÞ ¼  T U:

ð10Þ

Then the denominator on the right-hand side of eq. (6) is Z pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi d exp½H P ð; ; hÞ ¼ N =jUj: ð11Þ

2.3

Reconstruction as determining the mean of the posterior distribution The posterior distribution is given by Pðj; A; RÞPðj; hÞ

Pðj; A; Þ ¼ Z

ð12Þ

d Pðj; A; RÞPðj; hÞ ¼ where

1 exp½Hð; ; A; Þ; Z

ð13Þ

Z Z  PðjA; Þ ¼ d Pðj; A; RÞPðj; hÞ Z ¼ d exp½Hð; ; A; Þ;

ð14Þ ð15Þ

ð16Þ T

RA Þ Vð  V

1

T

RA Þ

V  U þ RA A:

ð17Þ ð18Þ

V is a positive definite symmetric matrix,   ð; h; RÞ is the set of hyperparameters, and H is the Hamiltonian of states at f; ; A; g. The smaller H the state f; ; A; g gives, the larger the posterior probability Pðj; A; Þ is. The partition function Z is given by sffiffiffiffiffiffiffiffiffiffiffiffiffi jUjRN T ðRR2 AV 1 AT Þ Z¼ : ð19Þ e N jVj Here we define the estimator of the original image as the mean of the posterior distribution: Z  ðÞ  hij; A;o ¼ d Pðj; A; Þ ð20Þ ffiffiffiffiffiffiffi r Z jVj T T T 1 1 ¼ d  eðV RA Þ VðV RA Þ ð21Þ N  ¼ V 1 RAT : ð22Þ Note that at the limit of no noise (R ! 1) the reconstruction matrix V 1 RAT in eq. (22) is equal to ðAT AÞ1 AT , the Moore–Penrose pseudoinverse. In this paper we use as a measure of reconstruction performance the mean squared error between the reconstructed image and original image: MSEð; ; A; Þ  3.



T

T

where J is a matrix relating the qth pixel to the rth pixel. Its elements are given by X X Jq;r ¼ 2dðqrÞ;o  ðqrÞ;  ðqrÞ; ; ð9Þ 

1

pffiffiffiffiffiffiffiffiffiffiffiffiffiffi N =RN

1 ð   ðÞ Þ2 : N

ð23Þ

Theory

3.1 Hyperparameter estimation It should be noted that for good reconstruction we need appropriate values for the hyperparameters , h, and R. We often have to estimate the hyperparameters from the observed image, however, because in realistic situations we have no explicit a priori information about them.10–14) Optimal hyperparameters can be obtained by selecting the ones that maximize Z.12,15) We can see from eq. (15) that Z, which is also called evidence, is the likelihood of filter and hyperparameters, PðjA; Þ, given by marginalizing Pð; jA; Þ over the original image . Maximizing this value is thus equivalent to maximum likelihood estimation of the filter and the hyperparameters. For computational convenience, instead of actually maximizing Z itself we minimize F   log Z  1 ¼  log jUj þ N log R  N log   log jVj 2 ð24Þ þ  T ðR  R2 AV 1 AT Þ: This F to be minimized, the negative logarithm of marginalized likelihood, is called free energy. In the present study, since we consider the case where the filter shape is known, we fix the filter A and optimize only the hyperparameters . The optimal hyperparameters satisfy the following saddle point conditions:

054803-2

J. Phys. Soc. Jpn., Vol. 77, No. 5

S. TAJIMA et al.

 @F 1 X 1 ¼ ðU Þq;r  ðV 1 Þq;r Jq;r þ ðV 1 RAT ÞT JðV 1 RAT Þ @ 2 q;r  1 X 1 T ¼ ðU Þq;r  ðV 1 Þq;r Jq;r þ  ðÞ J ðÞ ¼ 0; 2 q;r  @F 1 X 1 ¼ ðU Þq;r  ðV 1 Þq;r þ ðV 1 RAT ÞT ðV 1 RAT Þ @h 2 q;r  1 X 1 T ¼ ðU Þq;r  ðV 1 Þq;r þ  ðÞ  ðÞ ¼ 0; 2 q;r !  T   @F 1 N X 1 ¼  ðV Þq;r ðAT AÞq;r þ ðI  RAV 1 AT Þ ðI  RAV 1 AT Þ @R 2 R q;r ! 1 N X 1 T ¼  ðV Þq;r ðA AÞq;r þ ð  A ðÞ ÞT ð  A ðÞ Þ ¼ 0: 2 R q;r 3.2 Diagonalization with the Fourier transform Since the preceding formulations do not refer to any details of the matrix A, theoretically we can estimate hyperparameters and reconstruct original images under any real transformation matrix. Practically, however, computational costs must be taken into consideration. For N-pixel images, the number of elements of matrices such as U or V increases with OðN 2 Þ, explosively increasing the costs for calculating determinants or inverses. Fortunately, however, when the filters are translational-symmetric filters we can introduce diagonalization using the Fourier transform so that the number of terms appearing after the diagonalization are kept to the order of N.6,7) In the following we assume that A is a translationalsymmetric filter: Ao;r ¼ Aq;rþq :

ð31Þ

And we can see from eq. (9) that the matrix J is also translational-symmetric. For a d-dimensional image  each side of which is N 1=d pixels long, the discrete Fourier transform is given by   X 2i T ~k ¼ r exp  1=d k r ; ð32Þ N r   1X~ 2i T ð33Þ k exp 1=d k r ; r ¼ N k N where i is the unit imaginary number and k and r are vectors respectively denoting wave number and pixel location. Both of these vectors have elements at fn; n þ 1; n þ 2; . . . ; n  1g (n ¼ N 1=d =2) in each dimension. Let us introduce following notation:   X 2i T ~ Xk  Xo;r exp  1=d k r ð34Þ N r   X 2i T Xq;rþq exp  1=d k r ð8qÞ; ð35Þ ¼ N r where X is an arbitrary matrix satisfying the translationalsymmetric property, Xo;r ¼ Xq;rþq ð8qÞ, for which J, A, U, and V can be substituted. Through the above transforms, the reconstruction of the original image [eq. (22)] is reformulated in diagonalized form as

rðÞ

  1 X ðÞ 2i T ¼ ~ exp 1=d k r ; N k k N

ð25Þ ð26Þ ð27Þ ð28Þ

ð29Þ

ð30Þ

ð36Þ

RA~k ð37Þ ~k : V~k Similarly, the equations for hyperparameter estimation [eqs. (24)–(30)] are reformulated as X 1   F¼  log U~ k þ log R  log N  log V~k 2 k

1 R2 jA~k j2 2 R þ j~k j ; ð38Þ N V~k 2 3 @ 2 3 J~k 6 @ 7 6 7 7 X6 6 @ 7 1 6 7 6 7F ¼ 6 7 6 7 2 4 ~ Uk 5 6 @h 7 k 4 @ 5 R2 jA~k j2 @R

1 1 1 1 ðÞ 2    þ j~k j : ð39Þ 2 U~ k V~k N ~kðÞ ¼

Comparing the reformulated equations in this subsection with those in the former subsections, we can see that the Fourier diagonalization reduces the computational redundancy. 3.3 Lower bound of the free energy The diagonalization described in the previous subsection not only reduces computational costs but also facilitates theoretical analyses. We first confirm in this subsection that the estimated hyperparameters are equal to the actual hyperparameters and then in the next subsection analyze the performance of our reconstruction algorithm in terms of the mean squared error between the reconstructed and original images. For theoretical analysis, we can remove the dependencies on noise distribution and image content by marginalizing over whole configurations of pixel luminance values in an image.7) Suppose that original image  and observed image  are randomly generated from the distribution defined by eqs. (3) and (6), and that the hyperparameters are set to o ¼ ðo ; ho ; Ro Þ. Here the mean of free energy F is given by

054803-3

J. Phys. Soc. Jpn., Vol. 77, No. 5

hFi ¼ h log PðjA; ÞijA;o Z ¼  d PðjA; o Þ log PðjA; Þ

S. TAJIMA et al.

ð40Þ , ð41Þ

 o h ho ¼ o and ¼ o: R R R R

ð56Þ

¼ h log PðjA; o ÞijA;o   Z PðjA; o Þ o þ d PðjA;  Þ log ; ð42Þ PðjA; Þ where  is the estimate of the original hyperparameters o . The second term in eq. (42) is the Kullback–Leibler divergence of PðjA; Þ from PðjA; o Þ, which is not less than zero. Therefore, hFi is found to have the following lower bound:

Since, at least in the averaged analysis shown here, the conditions minimizing the mean free energy [eq. (50)] also minimize the mean squared error [eq. (56)], the hyperparameters estimated by minimizing the free-energy criterion will also minimize the mean squared error. For this reason, in the subsequent sections we use MSEo as a performance index in image reconstruction.

hFi  h log PðjA; o ÞijA;o  hF o i:

4.1 Laplacian-of-Gaussian filter Let us apply the present theory to a concrete example. We take a Mexican-hat shaped filter as an example of a translational-symmetric filter. A Mexican-hat shaped filter has a spatial weight distribution consisting, like that shown in Fig. 1(a), of a central region that contributes positively to the original pixel, a near surrounding region that contributes negatively, and a far surrounding region that makes almost no contribution [Fig. 1(b)]. Such a filter works as an edge detector, responding strongly to changes in surface luminance and not at all to completely uniform surface luminance16,17) [Figs. 1(c) and 1(d)], and its output models the responses of neurons in the retina or lateral geniculate nucleus.18–20) Reconstruction under a Mexican-hat filter is interpreted as an idealized demonstration of decoding the neuronal signals found in early visual systems and extracting the information in raw visual input. A Laplacian-ofGaussian filter (LGF) is often used for roughly modeling the shape of the Mexican hat:   krk2 2 Ao;r / r exp  2 2

2   2 krk krk2 ¼ 2  1 exp  2 : ð57Þ  2 2 2 When we normalize it so that X A2o;r ¼ 1; ð58Þ

ð43Þ

Equality is satisfied when   o . Writing eq. (42) in diagonalized form, we have 8 > > >  o o U~ R 1 X< hFi ¼ log k o þ 1 þ log N 2 k > V~k > : 0 ~ 2 ~ 319 Uk R Uk R > > > B V~ 6 V~ 7C= B k 6 k 7C ð44Þ þB o o  1  log6 o o 7C @ U~ k R 4 U~ k R 5A> > > ; V~ko V~ko  ~o o U R 1X  log k o þ 1 þ log N ¼ hF o i; ð45Þ 2 V~ k

k

V~ko  U~ ko þ Ro jA~k j2 ; U~ ko  o J~k þ ho :

ð46Þ ð47Þ

Here the condition for equality in eq. (45) is U~ k R U~ ko Ro ¼ ð8kÞ ð48Þ V~k V~ko jA~k j2 1 jA~k j2 1 , þ ¼ o þ ð8kÞ ð49Þ J~k þ h R  J~k þ ho Ro Except for some special cases, e.g., jA~k j2 ¼ 0 ð8kÞ, we can see that the lower bound of free energy is given only by the actual hyperparameters: ð; h; RÞ ¼ ðo ; ho ; Ro Þ:

ð50Þ

3.4 Lower bound of the reconstruction error Similar manipulations yield the following formula for the marginalized mean squared error: hMSEð; ; A; Þi;jA;o Z Z ¼ d d Pð; jA; o Þ MSEð; ; A; Þ ð51Þ ( )



2 X 1 Ro jA~k j2 R 2 U~ k U~ ko ¼ 1 þ  ð52Þ R Ro V~k U~ ko 2N V~ko k X 1   MSEo ; ð53Þ ~o 2N V k k where MSEo is the lower bound. The equality condition is U~ k U~ ko ð54Þ ¼ o ð8kÞ R R



 o h ho  o J~k þ  ¼ 0 ð8kÞ ð55Þ , R R R Ro

4.

Results

r

the complete form of filter is given by rffiffiffiffi

  1 2 krk2 krk2 Ao;r ðÞ ¼   1 exp  2 ;   2 2 2

ð59Þ

where  is the parameter that represents the spatial extension of the filter. Equation (59) is represented in Fourier-transformed form as pffiffiffiffiffiffi   42 2 3 kkk2 22  2 kkk2 ~ Ak ðÞ ¼  exp  : ð60Þ N N A LGF is a bandpass-like filter,21) and by changing  we can control which part of the frequency domain is amplified by the filter. 4.2 Hyperparameter estimation and image reconstruction We first ran an experiment using a 64  64-pixel grayvalued image and setting  equal to 2. Figures 2 and 3 respectively illustrate typical results of reconstruction and hyperparameter estimation. The upper row in Fig. 2 shows (a) an original image, (b) a filtered one, (c) an observed one, and (d) a reconstructed one. The original image in the upper

054803-4

J. Phys. Soc. Jpn., Vol. 77, No. 5

S. TAJIMA et al.

(a)

(b)

(c)

(d)

0.2

A0,q 0.1 0 40

qy

0

40

20

20 0

qx

Fig. 1. A Mexican-hat-shaped filter. (a) Spatial distribution of the filtering weight of a Laplacian-of-Gaussian filter. (b) Schematic illustration of the filter function. White, black, and gray regions respectively have positive (facilitative), negative (suppressive), and negligible effects on the center pixel. Comparing an original image (c) with the filtered image (d), one sees that filtering emphasizes the contours.

(a)

(b)

(c)

(d)

artificial image natural image Fig. 2. Demonstrations of filtering, observation, and reconstruction: (a) original images, (b) filtered images, (c) observed images, and (d) reconstructed original images. We demonstrate two cases: one with an artificial image (top row) and the other with a natural image (bottom row). The original artificial image was randomly generated from a prior distribution. For the artificial image the hyperparameters were set to ðo ; ho ; Ro Þ ¼ ð102 ; 101 ; 103 Þ, and for the natural image they were set to ðo ; ho ; Ro Þ ¼ ð102 ; 101 ; 102 Þ.

×104

2.5

F

R

1.5

4

×10 1.2

1

10 2

0.5 10 2

h 10

F

×10 2.5 104 2 1.5 102 1 100 β 0.5 10-2 10-2 100 102 104 6 10

F

2 4

1 10 0

0

2 10 -2 106 104 10 10

R

0

10

-2

0.8 10 -2

h

10 0

β

10 2

10 -2 10 4

Fig. 3. Transition of estimates of hyperparameters plotted on the landscape of the free energy. The artificial image sample used here is shown in the upper part of Fig. 2(a). The meshes shown in left, middle, and right panels respectively show the free energy as a function of ðR; Þ with h fixed to 101 , of ðh; RÞ with  fixed to 102 , and of ð; hÞ with R fixed to 103 . The estimated hyperparameters were initially set to ð; h; RÞ ¼ ð100 ; 100 ; 100 Þ, and one can see that the estimates converged near the actual values, ðo ; ho ; Ro Þ ¼ ð102 ; 101 ; 103 Þ. Image size N ¼ 256  256. The updating rate was 0:01=N.

row is an artificial image generated randomly from the prior distribution given by eq. (6), while the original image in the lower row is a natural image normalized to have zero-mean. The hyperparameters were estimated by 20,000 iterations of gradient descent according to eq. (39). Figure 3 shows the time evolution of the estimators for the hyperparameters in the case of the artificial image [upper part of Fig. 2(d)] and shows that they converged to almost the correct values. Close inspection of the result of reconstruction for the natural image [lower part of Fig. 2(d)] reveals that while the

contours were reconstructed fairly well, the surface luminances of pixels far from contours were not necessarily preserved. This failure is due to the observation noise. 4.3 The optimal filter parameter The analytical lower bound MSEo (eq. 53) gives a criterion for discussing filtering effects. Suppose that we want to transfer an image through a noisy communication channel. When we apply a LGF, which filter parameter is the best? Moreover, is the reconstruction obtained using a LGF

054803-5

J. Phys. Soc. Jpn., Vol. 77, No. 5

× 10

-3

Theoretical Simulated

MSE o

5 4 3

No filter

2

4

1 × 10

3

-3

6 γ

10

30

-1

R = 10-2 R = 10-0

R = 102

R = 10-2 R = 100

R = 104 R=

R = 102

106

R = 104, 106

1

3

6 γ

10

30

Fig. 5. Theoretical performance curves calculated for various noise levels. The stars mark the lowest point on each curve, and the dashed lines show the no-filter reconstruction performances corresponding to the curves with the same colors. N ¼ 256  256.

2 1 0

5 4 3 2 1 0

R = 1000

3 MSEo

-3 × 10

No filter

1

3

6 γ

10

30

30

30

Fig. 4. The reconstruction performance dependencies on the filter parameter . Each panel shows MSEo as a function of filter parameter  under a different noise condition with  ¼ 102 and h ¼ 105 . The smooth dark curves are derived from eq. (53), and the grey lines connect plots of the averages of actual calculations for 50 artificial images. The error bars represent standard deviations. The stars denote the lowest points on the theoretical curves. The dashed lines represent the performance obtained without filtering (i.e., where 8k; A~k ¼ 1). N ¼ 256  256.

better or worse than that obtained without using one? In other words, should we send the LGF-processed image or just send the original image? First, let us consider optimizing the filter width parameter  so that it minimizes MSEo . Here we consider the parameter range between 1 and 30 ( ¼ 1 corresponds to the pixel size and  ¼ 30 corresponds to roughly 1/10 of the length of each side of 256  256-pixels image,). Figure 4 shows the reconstruction error dependency on the filter parameter  under two different noise conditions: R ¼ 101 and 103 . The darker lines show the theoretically derived lower bound [MSEo , eq. (53)], the lighter lines show the means of 50 simulations with artificial images like the one shown in Fig. 2, and the error bars show the standard deviations. Here the filter parameter  is assumed to be known. We can see that the simulation results coincide with the theoretical prediction. This indicates that the results of the hyperparameter estimations gave the theoretical minimum reconstruction errors. The stars on the theoretical lines in the figure indicate theoretical minima, that is, the lowest points of the lower bounds giving the optimal  values for reconstruction. Comparing Figs. 4(a) and 4(b), we see that the optimal  is larger when R ¼ 1 than when R ¼ 1000. In other words, the broader LGF is preferred under the severer noise condition (note that the smaller R indicates more noise). Therefore, the optimal filter design can depend on values of the hyperparameters. We investigated this dependence over wider ranges of R. Figure 5 shows the plots of 21 theoretical lines (two of them are already shown in Fig. 4) obtained when the noise level was varied from R ¼ 102 to 106 . The solid and dashed lines respectively represent MSEo curves under LGF-applied and no-filter-applied conditions. The minima (stars) increasing

γ optimal

MSEo

6

(b)

LGF No filter

R=1

7

γ optimal

(a)

S. TAJIMA et al.

10 6 3 1

10 6 3 1

10-4 10-2 100 102 104 106

10-2

100

R

β

(a)

(b)

102

104

Fig. 6. Optimal filter parameter  for different noise and smoothness levels (N ¼ 256  256).

with decreasing R are consistent with the preceding paragraph: the larger  value is preferred in the severer noise condition. One sees in Fig. 6 that the optimal  is inversely correlated with R and positively correlated with the smoothness hyperparameter . Figure 4 also tells us that whether using a LGF is better than not using one depends on the hyperparameters and the filter parameter. Under a noisy condition [Fig. 4(a)] the errors near the optimal- condition are smaller than those in the no-filtering condition (dashed line), implying that the appropriate filtering can reduce the reconstruction errors. When the filter parameters are far from the optimal, however, the errors tend to be larger than those in the no-filtering condition. Thus, whether filtering is or is not beneficial depends on how well it is adjusted to the condition of the transferring channel. Figure 4(b), where the noise is much less than in Fig. 4(a), shows a situation in which the errors in the LGF-filtering condition are never smaller than those in the no-filtering condition. These results thus reveal that using a LGF before transferring an image can improve the reconstructing performances obtained when a communication channel is noisy if the filter parameter is optimal or nearly optimal. This is also seen in Fig. 5. When R > 1010 the errors under the LGF and no-filtering conditions are almost the same. As R decreases (i.e., as the channel becomes noisier) the errors increase under both conditions. Interestingly, however, the rate at which errors increase with increasing noise is slower in the LGF condition than in the no-filter condition. Figure 7 shows this more clearly. When the communication channel is noisier (smaller R) the LGF

054803-6

J. Phys. Soc. Jpn., Vol. 77, No. 5

× 10 -3

S. TAJIMA et al.

LGF is better

that if the noise in the channel is greater than the noise level at the critical point, we would get better reconstruction if we processed images with a LGF before sending them but if the channel noise is below this critical level we should just send raw images.

No filter is better

MSE omni

5 4 3

5.

R = 101.30

2 1

This section discusses the results of reconstructions with a natural image (Fig. 8) in order to help intuitively understand why the error depends on the hyperparameters and the filter parameter  and to show how the optimal filter values are decided. The original image shown in Fig. 8(a) was tested under different noise and filter conditions [Figs. 8(b)–8(m)] and the hyperparameters were estimated. The noise levels were set to R ¼ 100 (b)–(e), R ¼ 101:3 (f)–(i), and R ¼ 103 (j)–(m); and the filter parameter  was set to 2.5 (b), (f), and (j), 5.5 (c), (g), and (k), and 9 (d), (h), and (l). The corresponding reconstructions obtained when no filter was used are shown in parts (e), (i), and (m). The optimal filter parameter  was about 9 for R ¼ 100 , 5.5 for R ¼ 101:3 , and 2.5 for R ¼ 103 . Since, as shown in the previous section, R ¼ 101:3 is the critical noise level, at this noise level the

LGF No filter

0

-1 10-10 10-8 10-6 10-4 10-2 100 102 104 106 108 1010

R Fig. 7. Lower bound of mean squared reconstruction error (at optimal filter condition) versus noise level R (N ¼ 256  256).

causes smaller errors than not filtering does, and when the channel is less noisy (larger R) the LGF causes larger errors than not filtering does. There is therefore a critical point at a medium noise level (R ¼ 101:30 in this figure). This means (a)

Discussion

Original image

γ = 2.5

γ = 5.5

(Narrow filter)

γ =9

(Middle filter)

(c)

(b)

No filter

(Broad filter)

(d)

(e)

R = 100 (High noise level)

optimal (g)

(f)

(h)

(i)

R = 101.3 (Middle noise level)

optimal (k)

(j)

(l)

(m)

R = 103 (Low noise level)

(n)

(

1 0.5 A0,q 0 -0.5 -30 -20 -10

optimal)

0

qx

(o)

(p)

1 0.5 A 0,q 0 -0.5 10 20 30 -30 -20 -10

1 0.5 A0,q 0 -0.5 10 20 30 -30 -20 -10

pixels

0

qx

pixels

(q) 1

A0,q 0.5 0

qx

10 20 30

pixels

0 -0.5 -30 -20 -10

0

qx

10 20 30

pixels

Fig. 8. Reconstruction samples for a natural image under different filter and noise conditions. (a) the original image. (b)–(d), (f)–(h), and (j)–(l) show reconstructed images obtained with various filter sizes at high, medium, and low or noise levels. (e), (i), and (m) show reconstructed images obtained under no-filtering conditions. (n), (o), and (p) are lateral views of the filters at  ¼ 2, 5, and 7. (q) no filter (represented as a delta-function) (N ¼ 256  256).

054803-7

J. Phys. Soc. Jpn., Vol. 77, No. 5

S. TAJIMA et al.

error in the optimal-filter condition was roughly equal to that in the no-filtering condition. That is, under the medium noise level (f)–(h), the best reconstruction obtained when a LGF was used was obtained when one with  ¼ 5:5 (g) was used, and that reconstruction was as good as that obtained when no filter was used (i). Under the high noise level (b)–(d), the best reconstruction was obtained when a filter with  ¼ 9 was used. Under the low noise level (j)–(l), best reconstruction obtained when a LGF was used was obtained when one with  ¼ 2:5 was used, but that reconstruction was worse than the one obtained when no filter was used (m). Parts (n), (o), (p), and (q) of this figure show cross sections of the output distributions for each filter and for no filter Note that the no-filter condition is equivalent to one with a deltafunction-shaped filter. Of the four filter conditions, those with  ¼ 2:5 (the narrow filter) and with no filter were the ones most dependent on the noise level. Since the narrow filter did not emphasize properties in the low-frequency domain, it hardly preserved the surface luminance when the channel was very noisy. In addition, objects having no sharp edges (e.g., out-of-focus background objects) were almost unrecognizable in the reconstruction under the highest noise level. Not filtering was also fragile but gave a blurred reconstruction result under the same high noise level. The broad filter was surprisingly robust to noise. Although the precision of its reconstruction was not greater than that of the narrow filter or no filter, we can see that it preserved both surface luminance and abstract contour structures even under quite noisy conditions. Under the medium noise level, the moderately broad filter gave the best reconstruction. The optimal value of the filter parameter is thus determined by the trade-off between preserving edges and preserving surfaces. In this paper we present a simple case in which only a single filter is used and the observer knows the filter parameter. An interesting extension of the present model is to multi-filtering processes (either parallel or serial). Another extension would be to filter estimation, when we do not have any a priori information about the filter and have to estimate the filter parameters as well as the hyperprameters. 6.

Conclusion

We formulated computations for Bayesian image reconstruction from filtered images and for hyperparameter

estimation in a framework of free-energy minimization. Rewriting the calculations in Fourier-diagonalized forms, we proposed methods for reducing computational costs and for analytically evaluating the reconstruction and estimation performances. Experiments with Laplacian-of-Gaussian filters confirmed theoretical predictions and demonstrated that reconstruction performance depends on the hyperparameters and the filter parameter. They showed that a Laplacianof-Gaussian filter reduces the mean square error of the reconstruction when the channel is comparatively noisy and the filter is optimized for it. Acknowledgements This work was partially supported by Grants-in-Aid for Scientific Research on Priority Areas (Grant Nos. 18079003, 18020007, and 18079012), and a Grant-in-Aid for Scientific Research (C) (Grant No. 16500093) from the Ministry of Education, Culture, Sports, Science and Technology of Japan.

1) 2) 3) 4) 5) 6) 7) 8) 9) 10) 11) 12) 13) 14) 15) 16) 17) 18) 19) 20) 21)

054803-8

K. Tanaka: J. Phys. A 35 (2002) R81. J. Marroquin, S. Mitter, and T. Poggio: J. Am. Stat. Assoc. 82 (1987) 76. H. Nishimori and K. Y. M. Wong: Phys. Rev. E 60 (1999) 132. J. Inoue and D. M. Carlucci: Phys. Rev. E 64 (2001) 036121. A. S. Willsky: Proc. IEEE 90 (2002) 1396. K. Tanaka and T. Horiguti: Phys. Rev. E 65 (2002) 046142. J. Tsuzurugi and M. Okada: Phys. Rev. E 66 (2002) 066704. G. J. Burton and I. R. Moorhead: Appl. Opt. 26 (1987) 157. D. J. Field: J. Opt. Soc. Am. A 4 (1987) 2379. J. Besag: J. R. Stat. Soc. B 48 (1986) 259. S. Lakshmanan and H. Derin: IEEE Trans. Pattern Anal. Mach. Intell. 11 (1989) 799. J. M. Pryce and A. D. Bruce: J. Phys. A 28 (1995) 511. Z. Zhou, R. N. Leahy, and J. Qi: IEEE Trans. Image Process. 6 (1997) 844. R. Molina, A. K. Katsaggelos, and J. Mateos: IEEE Trans. Image Process. 8 (1999) 231. D. J. C. MacKay: Neural Comput. 4 (1992) 720. D. Marr and E. Hildreth: Proc. R. Soc. London, Ser. B 207 (1980) 187. V. Torre and T. A. Poggio: IEEE Trans. Pattern Anal. Mach. Intell. 8 (1986) 147. S. W. Kuffler: J. Neurophysiol. 16 (1953) 37. H. B. Barlow: J. Physiol. 119 (1953) 69. J. E. Dowling and B. B. Boycott: Proc. R. Soc. London, Ser. B 166 (1966) 80. D. Roberge and Y. Sheng: Appl. Opt. 33 (1994) 5287.

Bayesian-Optimal Image Reconstruction for ...

May 12, 2008 - marginalized likelihood, is called free energy. In the present ..... control which part of the frequency domain is amplified by the filter. .... 100. 102. 104. (a). (b). Fig. 6. Optimal filter parameter for different noise and smoothness.

303KB Sizes 1 Downloads 237 Views

Recommend Documents

Image Reconstruction in the Gigavision Camera
photon emission computed tomography. IEEE Transactions on Nuclear Science, 27:1137–1153, June 1980. [10] S. Kavadias, B. Dierickx, D. Scheffer, A. Alaerts,.

Image Processing and 3-D Reconstruction in Electron ...
the specimen in the TEM, different views can be obtained, which is the key for ...... smallpox, is one of the largest (∼360 nm) and most complex viruses whose ...

Image reconstruction of multi-channel photoacoustic ...
J.L.J.: E-mail: [email protected], Telephone: +64 (0)9 373 7599 ext 88747 ... The experiment was fully automated using the open-source PLACE ...

See the Difference: Direct Pre-Image Reconstruction ...
•We realize that the associated feature computation of HOG is piecewise differentiable and ... Visualizing Object Detection Features. In CVPR, 2013. ... histogram binning as spatial filtering. HOG vector v. [v1,v2,v3,v4,v5,v6,···] v. / kvk+∈ c

Image Reconstruction in the Gigavision Camera - Research at Google
to estimate the light intensity through binary observations. We model the ... standard memory chip technology is fast and has low cost. Different from the ..... The input image is 'building.bmp' .... counting imaging and its application. Advances in 

High-resolution image reconstruction from ... - Wiley Online Library
Mar 25, 2004 - Translated Low-Resolution Images with Multisensors. You-Wei Wen,1 ... high-resolution image reconstruction is turned into the problem of.

See the Difference: Direct Pre-Image Reconstruction ...
Wei-Chen Chiu and Mario Fritz. Max Planck Institute for Informatics, Saarbrücken, Germany. {walon, mfritz}@mpi-inf.mpg.de. Motivation. •HOG descriptor [3] has ...

mobilizing capacity for reconstruction and development - Human ...
4.9 Paying the Price of Conflict: A Strategic Challenge ...... which deals with governance, democracy and the rule ..... money monthly to a member of the club.

Groupwise Constrained Reconstruction for Subspace Clustering - ICML
dal, 2009; Liu et al., 2010; Wang et al., 2011). In this paper, we focus .... 2010), Robust Algebraic Segmentation (RAS) is pro- posed to handle the .... fi = det(Ci)− 1. 2 (xi C−1 i xi + νλ). − D+ν. 2. Ci = Hzi − αHxixi. Hk = ∑ j|zj =k

mobilizing capacity for reconstruction and development - Human ...
Population with Access to Education (%) .... broken promises of security and human development. ... access to knowledge, better nutrition and health services,.

Groupwise Constrained Reconstruction for Subspace Clustering
50. 100. 150. 200. 250. Number of Subspaces (Persons). l.h.s.. r.h.s. difference .... an illustration). ..... taining 2 subspaces, each of which contains 50 samples.

Groupwise Constrained Reconstruction for Subspace Clustering - ICML
k=1 dim(Sk). (1). Unfortunately, this assumption will be violated if there exist bases shared among the subspaces. For example, given three orthogonal bases, b1 ...

Groupwise Constrained Reconstruction for Subspace Clustering
The objective of the reconstruction based subspace clustering is to .... Kanade (1998); Kanatani (2001) approximate the data matrix with the ... Analysis (GPCA) (Vidal et al., 2005) fits the samples .... wji and wij could be either small or big.

PartBook for Image Parsing
effective in handling inter-class selectivity in object detec- tion tasks [8, 11, 22]. ... intra-class variations and other distracted regions from clut- ...... learning in computer vision, ECCV, 2004. ... super-vector coding of local image descripto

PartBook for Image Parsing
effective in handling inter-class selectivity in object detec- tion tasks [8, 11, 22]. ... automatically aligning real-world images of a generic cate- gory is still an open ...

RTTI reconstruction - GitHub
Mobile. Consumer. Cmd. Consumer. Munch. Sniffer. FileFinder. FileCollect. Driller ... o Custom data types: ✓ wrappers ... Identify Custom Type Operations ...

Methodology for 3D reconstruction of objects for ...
return high quality virtual objects. Based on this ... A better solution is the usage of a cloud of web service solution. A review of ... ARC3D (Vergauwen, 2006) is a free web service that provides a standalone software application for uploading.