IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 360-363

International Journal of Research in Information Technology (IJRIT)

www.ijrit.com

ISSN 2001-5569

Image Fusion With Undecimated Wavelet Transform Neethu K Department of Computer Science and Engineering Malabar College of Engineering and Technology Thrissur, Kerala, India [email protected]

Abstract—Image Fusion is the procedure of combining useful features from multiple sensor image inputs to a single composite. The resulting image will be more informative than any of the input images. Multiscale transforms are among the most popular techniques in the field of pixel-level image fusion. However, the fusion performance of these methods often deteriorates for images derived from different sensor modalities. In this paper, we demonstrate that for such images, results can be improved using a undecimated wavelet transform based fusion scheme, which splits the image decomposition process into two successive filtering operations using spectral factorization of the analysis filters. The actual fusion takes place after convolution with the first filter pair. Its significantly smaller support size leads to the minimization of the unwanted spreading of coefficient values around overlapping image singularities. The combination of these techniques leads to a fusion framework, which provides clear advantages over traditional multiscale fusion approaches, independent of the underlying fusion rule, and reduces unwanted side effects such as ringing artifacts in the fused reconstruction. Here we combine color images also, for that we are using discrete wavelet transform.

Index Terms-Image fusion, spectral factorization, undecimated wavelet transform (UWT), discrete wavelet transform(DWT). I.

INTRODUCTION

In computer vision, Image fusion is the process of combining relevant information from two or more images into a single image. The resulting image will be more informative than any of the input images. In remote sensing applications, the increasing availability of space borne sensors gives a motivation for different image fusion algorithms. Several situations in image processing require high spatial and high spectral resolution in a single image. Most of the available equipment is not capable of providing such data convincingly. Image fusion techniques allow the integration of different information sources. The fused image can have complementary spatial and spectral resolution characteristics. However, the standard image fusion techniques can distort the spectral information of the multispectral data while merging. Here using a novel undecimated wavelet transform based fusion scheme, which splits the image decomposition process into two successive filtering operations using spectral factorization. For the color images we used the DWT technique. The process of image fusion can be performed at pixel-, feature- or decision-level. Image fusion at pixel-level represents the combination of information at the lowest level, since each pixel in the fused image is determined by a set of pixels in the source images. Generally, pixel-level techniques can be divided into spatial and transform domain technique. Among the transform domain techniques, the most frequently used methods are based on multiscale transforms where fusion is performed on a number of different scales and orientations, independently. In multiscale pixel-level image fusion, a transform coefficient of an image is associated with a feature if its value is influenced by the feature’s pixel. In order to simplify the discussion, we will refer to a given decomposition level, orientation band and position of a coefficient as its localization. A given feature from one of the source images is only conserved correctly in the fused image if all associated coefficients are employed to generate the fused multiscale representation. However, in many situations this is not practical since, given a localization, the coefficient from image may be associated to a feature and the coefficient from image IB may be associated to a feature. In this case, choosing one coefficient instead of the other may result in the loss of an important salient feature from one of the source images. For example, in the case of a camouflaged person hiding behind a bush the person may appear only in the infrared image and the bush only in the visible image. If the bush has high textural content, this may result in large coefficient values at coincident localizations in both decompositions of

Neethu K,IJRIT

360

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 360-363

an infrared-visible image pair. However, in order to conserve as much as possible of the information from the scene, most coefficients belonging to the person (infrared image) and the bush (visible image) would have to be transferred to the fused decomposition. If there are many such coefficients at coincident localizations, a fusion rule that chooses just one of the coefficients for each localization may introduce discontinuities in the fused subband signals. These may lead to reconstruction errors such as ringing artifacts or substantial loss of information in the final fused image. It is important to note that the above mentioned problem is aggravated with the increase of the support of the filters used during the decomposition process. This results in an undesirable spreading of coefficient values over the neighborhood of salient features, introducing additional areas that exhibit coefficients in the source images with coincident localizations. In this paper, we propose a novel UWT-based pixel-level image fusion approach, which attempts to circumvent the coefficient spreading problem by splitting the image decomposition procedure into two successive filter operations using spectral factorization of the analysis filters.

Figure 1. Schematic diagram of the proposed framework

A schematic flow-chart of the suggested image fusion framework is given in Fig. 1. The co registered source images are first transformed to the UWT domain by using a very short filter pair, derived from the first spectral factor of the overall analysis filter bank. After the fusion of the high-pass coefficients, the second filter pair, consisting of all remaining spectral factors, is applied to the approximation and fused, detail images. This yields the first decomposition level of the proposed fusion approach. Next, the process is recursively applied to the approximation images until the desired decomposition depth is reached. After merging the approximation images at the coarsest scale the inverse transform is applied to the composite UWT representation, resulting in the final fused image. For the color images we used the DWT technique. The discrete wavelets transform allows the image decomposition in different kinds of coefficients preserving the image information. Such coefficients coming from different images can be appropriately combined to obtain new coefficients so that the information in the original images is collected appropriately. Once the coefficients are merged the final fused image is achieved through the inverse discrete wavelets transform. II.

MULTISCALE IMAGE FUSION

In general, pixel-level techniques can be divided into spatial and transform domain techniques. As for spatial domain techniques, the fusion is performed by combining all input images in a linear or non-linear fashion using weighted average, variance or total-variation based algorithms. Transform domain techniques map (transform) each source image into the transform domain (e.g. wavelet domain), where the actual fusion process takes place. The final fused image is obtained by taking the inverse transform of the composite representation. The main motivation behind moving to the transform domain is to work within a framework, where the image’s salient features are more clearly depicted than in the spatial domain. While many different transforms have been proposed for image fusion purposes, most of the transform domain techniques use multiscale transforms. This is motivated by the fact that images tend to present features in many different scales. In addition, the human visual system seems to exhibit high similarities with the properties of multiscale transforms. More precisely, strong evidence exists that the entire human visual field is covered by neurons that are selective to a limited range of orientations and spatial frequencies, and can detect local features like edges and lines. This makes them very similar to the basis functions of multiscale transforms A. Udecimated Wavelet Transform The UWT is implemented using a filter bank which decomposes an one-dimensional (1-D) signal c0 into a set W = {w1, . . . ,wJ , cJ }, in which wj represents the highpass or wavelet coefficients at scale j and cJ are the lowpass or approximation coefficients at the lowest scale J. The passage from one resolution to the next one is obtained using the “à

Neethu K,IJRIT

361

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 360-363

trous” algorithm, where the analysis lowpass and analysis high-pass filter h and g are upsampled when processing the jth scale, where j = 0, . . . , J − 1. Due to the nonsubsampled nature of the UWT, many ways exist to construct the fused image from its wavelet coefficients. For a given analysis filter bank (h, g), any synthesis filter bank( ˜h, ˜g) satisfying the perfect reconstruction condition of UWT can be used for reconstruction. The Isotropic Undecimated Wavelet Transform, which is frequently used in multispectral image fusion. In this approach, only one detail image for each scale is obtained and not three as in the general case. It is implemented using the non-orthogonal, 1-D filter bank. For example, due to the lack of convolutions during reconstruction no additional distortions are introduced when constructing the fused image. Furthermore, since the fused image can be obtained by a simple co-addition of all detail images and the approximation image, a very fast reconstruction is possible. On the other hand, distortions introduced during the fusion process remain unfiltered in the reconstructed image. III.

UWT-BASED FUSION SCHEME WITH SPECTRAL FACTORIZATION

An input image can be represented in the transform domain by a sequence of detail images at different scales and orientations along with an approximation image at the coarsest scale. The spectral factorization method proposed here can be employed together with any fusion rule. Therefore, in order to assess the effectiveness of the proposed method, we applied four different fusion rules. The first investigated combination scheme is the simple “choose max” or maximum selection fusion rule. By this rule the coefficient yielding the highest energy is directly transferred to the fused decomposed representation. The simple choose max fusion rule does not take into account that, by construction, each coefficient within a multiscale decomposition is related to a set of coefficients in other orientation bands and decomposition levels. Since the combination schemes of fusion1 and fusion 2 suffer from a relative low tolerance against noise which may lead to a “salt and pepper” appearance of the selection maps, robustness can be added to the fusion process using an area based selection criteria “Multisensor pixel-level image fusion” . IV.

DISCRETE WAVELET TRANSFOM

The discrete wavelets transform allows the image decomposition in different kinds of coefficients preserving the image information. Such coefficients coming from different images can be appropriately combined to obtain new coefficients so that the information in the original images is collected appropriately. Once the coefficients are merged the final fused image is achieved through the inverse discrete wavelets transform, where the information in the merged coefficients is also preserved. The Discrete Wavelet Transform, which is based on sub-band coding is found to yield a fast computation of Wavelet Transform. It is easy to implement and reduces the computation time and resources required. In the DWT based image fusion, the key step is to define a fusion rule to create a new composite multiresolution representation. Up to now, the widely used fusion rule is the maximum selection scheme. This simple scheme just selects the largest absolute wavelet coefficient at each location from the input images as the coefficient at the location in the fused image. DWT was developed to apply the wavelet transform to the digital world. Filter banks are used to approximate the behavior of the continuous wavelet transform. In this paper we are including color images also. For this we use the DWT technique. . Color image fusion is the process of integrating one or more color images to enhance clarity of the image. As color images are best suited for representation and analysis, and these images may be corrupted by addition of noise by any means, that may be caused by an imaging system or may due to effects of environment. So, noise must be eliminated for better representation and analysis. The application of color image fusion simplifies object identification and helps in better cognition can be applied in robot vision, image classification, concealed weapon detection. The fusion of images is performed on images acquired from different instrument modalities. The Fusion produces a single image from the set of input images. The fused image should have wholesome information useful for human and machine perception. Image fusion improves reliability and capability. Fusion of color images can be carried by first converting color images into gray images and apply fusion for the gray images and then convert fused grey image back to color image. The use of color in image fusion is motivated by: Color is important in object recognition and human eyes can discern thousands of colors. The Discrete Wavelet Transform yields a fast computation of Wavelet Transform. It is easy to implement and reduces the computation time and resources required. The discrete wavelet transform was developed to apply the wavelet transform to the digital world. The signal is decomposed with a high-pass filter and a low-pass filter. The coefficients of these filters are computed using mathematical analysis.

Neethu K,IJRIT

362

IJRIT International Journal of Research in Information Technology, Volume 2, Issue 6, June 2014, Pg: 360-363

V.

CONCLUSION

An image fusion approach is presented here. Image fusion is the process of integration of two or more source images into a single fused image. The resultant fused image is used to retain the important features of the source images. Color image fusion is the process of integrating one or more color images to enhance clarity of the image. The method we proposed here successfully improves fusion results. Our method spectrally divides the analysis filter pair into two factors which are then separately applied to the input image pair, splitting the image decomposition procedure into two successive filter operations. The actual fusion step takes place after convolution with the first filter pair. For the color images we used the DWT technique. The discrete wavelets transform allows the image decomposition in different kinds of coefficients preserving the image information. Such coefficients coming from different images can be appropriately combined to obtain new coefficients so that the information in the original images is collected appropriately. Once the coefficients are merged the final fused image is achieved through the inverse discrete wavelets transform. . REFERENCES [1] [2] [3] [4] [5] [6] [7] [8] [9]

V. S. Petrovic and C. S. Xydeas, “Gradient-based multiresolution imagefusion,” IEEE Trans. Image Process., vol. 13, no. 2, pp. 228–237, Feb. 2004. Z. Zhang and R. S. Blum, “A categorization of multiscale decomposition-based image fusion schemes with a performance studyfor a digital camera application,” Proc. IEEE, vol. 87, no. 8, pp. 1315 1326, Aug. 1999. J. Nunez, X. Otazu, O. Fors, A. Prades, V. Pala, and R. Arbiol, “Multiresolution-based image fusion with additive wavelet decomposition,” IEEE Trans. Geosci. Remote Sens., vol. 37, no. 3, pp. 1204–1211, May 1999. M. Kumar and S. Dass, “A total variation-based algorithm for pixellevel image fusion,” IEEE Trans. Image Process., vol. 18, no. 9, pp. 2137–2143, Sep. 2009. Wirat Rattanapitak and Somkait Udomhunsakul,“Comparative Efficiency of Color Models for Multi-focus Color Image Fusion”,proc. Of international multiconference,march 17-19,2010 Shutao Li and Jianwen Hu“Image Fusion with Guided Filtering”, IEEE Trans. Image Process., vol. 22, no. 7, july. 2013. J. H. Jang and J. B. Ra,”Pseudo-Color Image Fusion Based on Intensity-Hue-Saturation Color Space”, Proc. IEEE international conference, August 20 - 22, 2008. S. Li and B. Yang, “Hybrid multiresolution method for multisensor multimodal image fusion,” IEEE Sensors J., vol. 10, no. 9, pp. 1519–1526, Sep. 2010. E. Lallier and M. Farooq, “A real time pixel-level based image fusion via adaptive weight averaging,” in Proc. 3rd Int. Conf. Inf. Fusion, vol. 2. Jul. 2000, pp. WeC3/3–WeC3/13.

Neethu K,IJRIT

363

Image Fusion With Undecimated Wavelet Transform

The process of image fusion can be performed at pixel-, feature- or decision-level. Image fusion at pixel-level represents the combination of information at the lowest level, since each pixel in the fused image is determined by a set of pixels in the source images. Generally, pixel-level techniques can be divided into spatial ...

75KB Sizes 1 Downloads 252 Views

Recommend Documents

Image Retrieval Based on Wavelet Transform and Neural Network ...
The best efficiency of 88% was obtained with the third method. Key Words: .... Daubechies wavelets are widely used in signal processing. For an ..... Several procedures to train a neural network have been proposed in the literature. Among ...

Approximation Algorithms for Wavelet Transform ... - CIS @ UPenn
Call this strip of nodes S. Note that |S| ≤ 2q log n. The nodes in S break A into two smaller subproblems L .... Proc. of VLDB Conference, pages 19–30, 2003.

Multispectral Multifocus Image Fusion with Guided ...
In this paper, we propose a novel solution to the above problem using guided .... Analytic image provides an easy solution for recov- ering local phase from the ...

Multi-sensor Fusion system Using Wavelet Based ...
Multi-sensor Fusion system Using Wavelet Based Detection. Algorithm Applied to Network Monitoring. V Alarcon-Aquino and J A Barria. Communications and ...

Wavelet Transform-based Clustering of Spectra in ...
Various kinds of statistical and mathematical data analysis techniques ... originating from different chemical applications and acquired by different spectrometers.

co-channel speech detection based on wavelet transform
Usable Speech Detection Using Linear Predictive Analysis –. A Model-Based Approach. Nithya Sundaram, Robert E. Yantorno, Brett Y. Smolenski and Ananth ...

Review of Image Fusion Techniques
IJRIT International Journal of Research in Information Technology, Volume 2, Issue 1, ... 1 M-Tech student, Electronics and Telecom, Government college of Engineering ..... which provides best approximation in image details of fused images.

CONTINUOUS WAVELET TRANSFORM Notes 1.pdf
6.3 Continuous wavelet transforms. In this section and the next we shall describe the concept of a continuous. wavelet transform (CWT), and how it can be approximated in a discrete form. using a computer. We begin our discussion by describing one typ

Adaptive directional wavelet transform using pre ... - IEEE Xplore
ADAPTIVE DIRECTIONAL WAVELET TRANSFORM USING PRE-DIRECTIONAL. FILTERING. Yuichi Tanaka. 1. , Madoka Hasegawa. 1. , Shigeo Kato. 1.

Approximation Algorithms for Wavelet Transform Coding of Data ...
Page 1 ... are applicable in the one-pass sublinear-space data stream model of computation. ... cinct synopses of data allowing us to answer queries approxi-.

Direction Scalability of Adaptive Directional Wavelet Transform: An ...
ternative of the traditional 2-D wavelet transform for image coding. It is able to ... Our method is effective for very low bitrate image coding, and is comparable to ...

Direction Scalability of Adaptive Directional Wavelet Transform: An ...
In image and video processing using wavelet transform (WT), multiresolution decomposition is one of the most important features. [1]–[3]. It represents an image by several multiresolution subbands. Since most images have higher energy in low-freque

image denoising using wavelet embedded anisotropic ...
*Network Systems & Technologies Ltd (NeST), Technopark Campus, Trivandrum INDIA, Email ... subband adaptive systems have superior performance, such as.

Adaptive compressed image sensing based on wavelet ...
Thus, the measurement vector y , is composed of dot-products of the digital image x with pseudo-random masks. At the core of the decoding process, that takes.

medical image fusion using cross scale coeffcient
decomposition-based fusion of volumetric medical images taking into account both intrascale and interscale consistencies. An optimal set ..... navigation, but also offers them the flexibility of combining modalities of their choice, which is importan

Investigation on image fusion of remotely sensed ...
In this paper the investigation on image fusion applied to data with significantly different spectral properties is presented. The necessity of placing high emphasis on the underscored problem is demonstrated with the use of simulated and real data.

Image-based fusion for video enhancement of night ...
School of computer science and engineering, Chengdu,. China, 611731. bShanghai Jiao Tong University, Institute of Image. Communications and Information ... online Dec. 29, 2010. .... where FL(x, y) is the final illumination image, M(x, y) is.