SALIENCY DETECTION VIA FOREGROUND RENDERING AND BACKGROUND EXCLUSION Yijun Li, Keren Fu, Lei Zhou, Yu Qiao, Jie Yang∗ Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong University, P.R.China ABSTRACT In this paper, a novel approach for image visual saliency detection is proposed from both the salient object (foreground) and the background perspective. To better highlight the salient object, we start from what is a salient object and adopt priors including contrast prior and center prior to measure the dissimilarity between different image elements. To better suppress the background, we focus on what is the background and measure the pixel-wise saliency by the minimum seam cost where the seam is an optimal 8-connected path from the pixel to some boundary pixel. The final saliency map is obtained by the combination of two measure systems which leads to the goal of both highlighting the salient object and suppressing the background. Both qualitative and quantitative experiments conducted on a benchmark dataset show that our approach outperforms seven state-of-the-art methods. Index Terms— Saliency detection, SODM map, BDM map, Saliency map 1. INTRODUCTION To accurately identify the most visually conspicuous object in the scene, known as the salient object, is very easy and fast for humans. However, it is quite challenging for a computer to match such a human behavior. If it succeeds in a cyber system, which means that the salient region can be automatically estimated, it will gain high values at many aspects, like isolating the object from cluttered background for surveillance [1, 2, 3] or allocating computational resources for subsequent processing. In general, the main goal of saliency detection lies in two directions: one is to predict a few eye fixation points in an image [4] and another one is to highlight the entire salient object region [5, 6, 7, 8, 9]. Our approach aims at the second goal and produces the object-level saliency map. For saliency detection, it is natural to begin with the description about what a salient object is and hence different prior knowledge is characterized from different perspectives. The most widely used prior is the feature contrast between different image elements. Itti [10] et al. extract the centersurround difference across multi-scale image features including luminance, color and orientation to find the prominent re∗ Corresponding

author: Jie Yang, [email protected]

(a) Input image

(e) Final saliency map

(b) SODM map

(c) BDM map

(d) Combination

Fig. 1: The framework of our approach gion. However, such local contrast only stands out at object boundaries. So region contrast based methods [5, 9] compute the global contrast as saliency which can usually highlight the entire object. Achanta [11] et al. propose a frequency-tuned method that directly defines pixel saliency using color difference from the image average color. Other priors like center prior [6, 9] or shape prior [12] based on a salient objects characteristics are also incorporated to help obtain better detection results. These approaches show superiority on the highlighting of the salient region. Actually, standing on the opposite side, it is quite creative to consider what the background is and hence pop out the salient object passively. Wei et al. [8] firstly propose a boundary prior, pointing to the fact that boundary pixels are almost background since photographers seldom put the salient object along image frame. The saliency of a pixel is measured by its shortest geodesic distance to the boundary component set. Actually the boundary prior and the center prior are to some extent coherent as the center prior assumes that the salient object is often framed near the center of the image. The main advantage of the boundary driven measure is its better performance on background suppression. However, simply starting one of the aforementioned two perspectives brings about two main problems: (i) It seldom realizes the highlighting of the salient object and the background suppression simultaneously. (ii) It may fail in some cases when images contain complex scenes. Therefore, we propose a novel approach from both the salient object and the background perspective by taking advantages of both and exploiting their mutual complementary

effect to solve mentioned problems. The main contributions of our approach lie in three aspects: (i) From the salient object angle, the saliency is measured in a region-wise way to incorporate several priors together by exploiting the feature dissimilarity between different regions. (ii) From the background angle, the pixel-wise saliency is defined as the minimum seam cost where the seam is an optimal 8-connected path from the pixel to some boundary pixel. The optimality is defined by an image energy function. (iii) The final saliency map is obtained by the combination of two measure systems and a smoothing strategy. The remainder of this paper is organized as follows. Section 2 introduces the proposed approach in detail. Section 3 conducts experiments to compare the proposed approach with several existing counterparts and results are also analyzed. The conclusion is given in Section 4.

(a) Input image

(b) Convex hull

(c) SODM map

Fig. 2: Exemplar results of the SODM map we select the top N superpixels which have the largest margin to the average of the image in LAB color space. This is different from [15] which selects interesting points at the cost of high computational time. A convex hull is used to bound these N superpixels. Then the center pc is computed as the average position of pixels within the hull. Given the center pc , the saliency value of ri is refined as:

2. METHODOLOGY

Sct (ri ) = e−α2 kpi −pc k2 × Scon (ri )

In the proposed method, two saliency measures, i.e., the salient object driven measure (SODM) and the background driven measure (BDM), are respectively designed given the input mage. Then a combination strategy is proposed to generate the final saliency map. Fig. 1 shows the resulting images in each step. Details are given in the following subsections.

Moreover, according to Gestalt laws [16], visual system tends to group similar regions together which implies that regions with similar colors ought to have closer saliency values. Hence we adopt a smoothing strategy as: 1 X −β1 kci −cj k2 e × Sct (rj ) (3) Sc (ri ) = Z1 j

2.1. The SODM Map

P where Z1 = j e−β1 kci −cj k2 serves as a normalization term. Fig. 2 illustrates a few exemplar results of the convex hull and the SODM map. It can be seen that though the salient object is framed off the image center, our hull still successfully encircles it.

Before computing the SODM map, the input image is partitioned into a number of superpixels [13] which have homogenous color components within. To begin with what a salient object is, we adopt the widely used prior, i.e., the contrast prior, which means that the color components belonging to a salient object often have strong contrast to their surroundings. Based on this general prior, the saliency value of superpixel ri is defined as: X Scon (ri ) = e−α1 kpi −pj k2 × kci − cj k2 (1) j

where ci , cj are the average color vectors in LAB color space as LAB better characterizes the human visual perception, and pi , pj are the average position vectors. It can be seen that it is a global way since for a superpixel ri , we compute its dissimilarity to all the rest superpixels. The term e−α1 kpi −pj k2 is the spatial constrain, which enhances the effect of nearer neighbors. Then we take the center prior into consideration. Different from [6, 12] which directly regard the image center as where the salient object often locates, our center is shifted to be chosen as the center of a convex hull which has high probability of covering the salient object. The hull is generated by searching for some potential salient regions. As validated in [11, 14], the salient part often highly deviates from the average of the image in feature space. Therefore

(2)

2.2. The BDM Map It is observed in Fig. 2 that while the salient object is highlighted, the SODM map is limited in background suppression. Therefore our attention is shifted from what is a salient object to what is the background. We exploit the boundary prior by treating boundary pixels as the background, and the following work is mainly inspired by [17]. In [17], it presents a classical method called seam carving which targets on resizing the content-aware image by searching a vertical (horizontal) seam from the top (left) to bottom (right) boundary which possesses the minimum seam cost. Instead of searching a seam for an image in [17], we aim at finding a seam for every pixel, starting from certain boundary pixel to the pixel itself. To minimize the cost, an energy function Ψ should be defined beforehand and here Ψ is defined by the gradient image of the SODM map by Sobel operator as follows: Ψ=|

∂Sc ∂Sc |+| | ∂x ∂y

(4)

Next we take the top boundary for example and elaborate for each pixel how to find its corresponding seam. With the

1 0.9 0.8 (a) Input image

(b) Energy function

0.7

(c) BDM map

0.6 Precision

(d) The last row of (b)

0.5 CA IT FT HC RC SF GS Ours

0.4 0.3 (e) Top S1

(f) Bottom S2

(g) Left S3

(h) Right S4

0.2

Fig. 3: Exemplar results of the BDM map

0.1 0

seam which is an 8-connected path of pixels and its corresponding seam cost similarly defined in [17], we search for the optimal seam that minimizes the seam cost in the following way: to traverse the image from the second row to the last row and compute the cumulative minimum energy S1 for all possible connected seams for each entry (i, j): S1 (i, j) = Ψ(i, j) + min(S1 (i − 1, j − 1), S1 (i − 1, j), S1 (i − 1, j + 1))

(5)

where the first row of S1 is equal to Ψ (i is for row and j is for column). Since S1 (i, j) is obtained in a cumulative way, S1 (i, j) is the minimum seam cost from certain top boundary pixel to the pixel (i, j). The fundamental principle behind our algorithm is that as the gradient at the edge of the salient object is high, so the seam cost of a salient object pixel will be leveled up when its path must travel across the edge and finally reach the image boundary. Likewise, the seam cost of a pixel to the bottom, left and right boundary can be found in the same way and hence we get other three maps S2 , S3 and S4 . The final BDM map is obtained as follows: Sbac (i, j) = min(S1 (i, j), S2 (i, j), S3 (i, j), S4 (i, j)) (6) For every pixel, we choose the minimum of its corresponding four seam cost as the final saliency. As can be seen in S1 (Fig. 3(e)), pixels faraway from the top boundary will get high seam cost values under two conditions: (i) The long distance causes the cumulative effect; (ii) The 8-connected path sometimes unavoidably travels across some strong gradient point. Therefore exploiting the minimum of four cost values can alleviate such a problem. We do not select the direct multiplication of four maps to avoid the over-suppression of the salient object. Fig. 3 presents detailed stepping results of the BDM map. It can be seen that by Eq. (5), as long as the edge of the salient object is highlighted, the whole object will be popped out. Fig. 3 also illustrates that our approach is robust to the image where the salient object is cropped on the boundary, which violates the boundary prior. In Fig. 3, the bird is partly cropped on the bottom boundary, but S2 is not seriously affected which is due to the energy function Ψ. Fig. 3(d) shows

0

0.2

0.4

0.6

0.8

1

Recall

Fig. 4: Precision-Recall curves comparisons for the proposed method and seven existing method. the last row of Fig. 3(b) (the red rectangle), which tells that the energy function has already enabled bottom boundary pixels on the bird (the middle bright part) to be more salient than boundary pixels in the background. Through the cumulative computation in Eq. (5), the bird body will get relatively high saliency values. If the bottom boundary is directly treated as the background, the saliency value of the whole bird body will be close to zero and hence missed in visual saliency detection. 2.3. Final Saliency Assignment Based on two aforementioned saliency measures, we can derive the saliency map as follows: Sal(i, j) = Sc (i, j) × Sbac (i, j)

(7)

where Sal is normalized to the range [0,1] afterwards. As shown in Fig. 1(d), the combination step further plays an influencing role on the background suppression. However, such a non-linear combination suppresses the background well but may still cause highlighting the entire salient object failed. Again we adopt a smoothing strategy as follows: Smap (ri ) =

1 X −β2 kSal(ri )−Sal(rj )k2 e × Sal(rj ) (8) Z2 j

where Sal( ri ), Sal(rj ) are the average saliency of the superP pixel ri , rj and Z2 = j e−β2 kSal(ri )−Sal(rj )k2 also serves as a normalization term. After smoothing, Fig. 1(e) illustrates a more satisfying result on highlighting the salient object while keeping the background suppressed. 3. EXPERIMENT AND COMPARISON The proposed approach for saliency detection is evaluated on the benchmark ASD dataset [11] which includes 1000 images

1 0.9

CA

0.8

IT

Precision

0.7 0.6

FT

0.5 0.4

SODM only BDM only Combination

0.3

HC

0.2 0.1

RC 0

0.2

0.4

0.6

0.8

1

Recall

Fig. 5: Precision-Recall curves of step effect comparisons of our algorithm. and their manually labeled ground truth. In the following experiments, to generate the SODM map, each image is presegmented into about K = 200 superpixels and other five parameters are all empirically chosen as: N = K × 20%, α1 = α2 = 0.005 and β1 = β2 = 0.001. Results of our method are compared with seven state-ofart counterparts including CA [6], IT [10], FT [11], HC [5], RC [5], SF [7] and GS [8]. Among them, the former six are standing on the point of what is the salient object and the last one GS [8] standing on the opposite side, i.e., what is the background. We mainly adopt the criteria introduced in [11] to evaluate the performance using precision recall (PR) curves. The averaged PR curves are obtained by binarizing the saliency map using thresholds ranging from 0 to 255. Intuitively, the PR curve represents the robustness of the algorithm in a crossimage fashion. Fig. 4 presents the PR curves of all methods and it is observed that the proposed method reaches the best. Our maximum precision rate is 95.5% with nearly 5% improvement. Specifically, when the recall is low (corresponding to the high threshold), the superiority of our approach illustrates that the salient object is better highlighted. When the recall is high (corresponding to the low threshold), our PR curve is still above others which implies that our method achieves better background suppression effect. Fig. 6 shows exemplar images and their corresponding saliency maps from various algorithms. Fig. 5 presents the evaluation for individual phase of our algorithm, respectively including SODM only, BDM only and the combination of both. Obviously, the SODM works better at low recalls and the BDM shows its superiority at high recalls. Our combination of both illustrates the complementary effect between SODM and BDM. Especially the conspicuous enhancement at high recalls (corresponding to the low threshold) tells that the background is being too suppressed to be over the low threshold, which is in consistent with what we

SF

GS

Ours

GT

Fig. 6: Visual performance comparisons of the proposed method and the seven existing methods. From top to down: Input image, CA, IT, FT, HC, RC, SF, GS, Ours and Ground Truth. desire for the background. Both quantitative and qualitative results show the benefit of combining the two measures and validates our idea that fully considering both the characteristics of the salient object and the background help gain greater saliency detection results. 4. CONCLUSION This paper proposes a novel approach for image visual saliency detection by fully exploiting the characteristics of both the salient object and the background. Starting from what is a salient object, we adopt the priors including contrast prior and center prior to measure the dissimilarity between different image elements. Moreover, by analyzing what is the background, we measure the pixel-wise saliency as the minimal seam cost based on boundaries. The final saliency map is obtained by the combination of two measure systems which leads to the goal of both highlighting the salient object and suppressing background. Both evaluation and comparison results show the effectiveness of the proposed method. Acknowledgements. We thank the anonymous reviewers for their valuable suggestions. This research is partly supported by NSFC, China (No: 61273258, 61375048), Ph.D. Programs Foundation of Ministry of Education of China (No.20120073110018).

5. REFERENCES [1] J. Han, K. Ngan, M. Li, and H. Zhang, “Unsupervised extraction of visual attention objects in color images,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 16, no. 1, pp. 141–145, 2006. [2] B. Ko and J. Nam, “Object-of-interest image segmentation based on human attention and semantic region clustering,” JOSA A, vol. 23, no. 10, pp. 2462–2470, 2006. [3] M. Donoser, M. Urschler, M. Hirzer, and H. Bischof, “Saliency driven total variation segmentation,” in proc. ICCV. IEEE, 2009, pp. 817–824. [4] X. Hou and L. Zhang, “Saliency detection: A spectral residual approach,” in proc. CVPR, 2007, pp. 1–8. [5] M. Cheng, G. Zhang, N. Mitra, X. Huang, and S. Hu, “Global contrast based salient region detection,” in proc. CVPR, 2011, pp. 409–416. [6] S. Goferman, L. Zelnik-Manor, and A. Tal, “Contextaware saliency detection,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 34, no. 10, pp. 1915–1926, 2012. [7] F. Perazzi, P. Krahenbuhl, Y. Pritch, and A. Hornung, “Saliency filters: Contrast based filtering for salient region detection,” in proc. CVPR, 2012, pp. 733–740. [8] Y. Wei, F. Wen, W. Zhu, and J. Sun, “Geodesic saliency using background priors,” in proc. ECCV, 2012, pp. 29– 42. [9] K. Fu, C. Gong, J. Yang, and Y. Zhou, “Salient object detection via color contrast and color distribution,” in proc. ACCV, 2012, pp. 111–122. [10] L. Itti, C. Koch, and E. Niebur, “A model of saliencybased visual attention for rapid scene analysis,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 20, pp. 1254–1259, 1998. [11] R. Achanta, S. Hemami, F. Estrada, and S. Susstrunk, “Frequency-tuned salient region detection,” in proc. CVPR, 2009, pp. 1597–1604. [12] H. Jiang, J. Wang, Z. Yuan, T. Liu, N. Zheng, and S. Li, “Automatic salient object segmentation based on context and shape prior.,” in proc. BMVC, 2011, pp. 7–18. [13] R. Achanta, A. Shaji, K. Smith, A. Lucchi, P. Fua, and ´ S. S¨usstrunk, “Slic superpixels,” Ecole Polytechnique F´ed´eral de Lausssanne (EPFL), Tech. Rep, vol. 149300, 2010. [14] R. Margolin, A. Tal, and L. Zelnik-Manor, “What makes a patch distinct?,” in proc. CVPR, 2013.

[15] C. Yang, L. Zhang, and H. Lu, “Graph-regularized saliency detection with convex-hull-based center prior,” IEEE Signal Processing Letters, vol. 20, no. 7, pp. 637– 640, 2013. [16] K. Koffka, “Principles of gestalt psychology,” pp. 2–4, 1935. [17] S. Avidan and A. Shamir, “Seam carving for contentaware image resizing,” in ACM Transactions on graphics (TOG), 2007.

Saliency Detection via Foreground Rendering and Background ...

Saliency Detection via Foreground Rendering and Background Exclusion.pdf. Saliency Detection via Foreground Rendering and Background Exclusion.pdf.

1MB Sizes 4 Downloads 262 Views

Recommend Documents

Foreground-Background Regions Guided ... - Semantic Scholar
ent values and found that k = 0.34 gives the best results, but Sauvola[2] ... The suitable range for θ is from -45 to 45 degrees. From these ranges, a set of filters is generated for different .... International Conference on Computer Analysis of Im

separating a foreground singer from background music
The magnitude spectrogram for a signal is a two-dimensional data structure ... a single instance of (t, f); rather a draw of a large number Q of quanta of (t, f) will ..... sub graph when the first-level latent variable selected is the voice. None of

Hierarchical Method for Foreground Detection Using Codebook Model
Jun 3, 2011 - important issue to extract foreground object for further analysis, such as human motion analysis. .... this object shall become a part of the background model. For this, the short-term information is employed ...... processing-speed-ori

A Saliency Detection Model Using Low-Level Features ...
The authors are with the School of Computer Engineering, Nanyang ... maps: 1) global non-linear normalization (being simple and suit- ...... Nevrez İmamoğlu received the B.E. degree in Com- ... a research associate for 1 year during 2010–2011 ...

Saliency Detection based on Extended Boundary Prior with Foci of ...
Page 1 of 5. SALIENCY DETECTION BASED ON EXTENDED BOUNDARY. PRIOR WITH FOCI OF ATTENTION. Yijun Li1. , Keren Fu1. , Lei Zhou1. , Yu Qiao1. , Jie Yang1∗. , and Bai Li2. 1. Institute of Image Processing and Pattern Recognition, Shanghai Jiao Tong Uni

Regional Principal Color Based Saliency Detection - Semantic Scholar
Nov 7, 2014 - School of Computer Science and Engineering, Nanjing University of Science and ... simple computation of color features and two categories of spatial ... In recent years, numerous ... a higher degree of importance [19,20].

Nonparametric Bottom-Up Saliency Detection Using ...
Dec 1, 2011 - Experimental results on two public saliency detection datasets show that our approach performs better than four state-of-the art ... Fourier spectral residual analysis algorithm for static im- age saliency. In this method, amplitude ...

Improved Saliency Detection Based on Superpixel ...
Categories and Subject Descriptors. I.4.8 [Image Processing and Computer Vision]: Scene. Analysis. General Terms. Algorithms, Experimentation, Performance.

Novelty-based Spatiotemporal Saliency Detection for ... - IEEE Xplore
Feb 15, 2016 - Abstract—The automated analysis of video captured from a first-person perspective has gained increased interest since the advent of marketed ...

Saliency Detection based on Extended Boundary Prior with Foci of ...
K) and its mean position and mean color in. LAB color space are denoted as pi and ci respectively (both. normalized to the range [0,1]). 2.1. Graph Construction.

Early Stage Botnet Detection and Containment via ... - Semantic Scholar
Detecting Botnet Membership with DNSBL Counterintelligence, pp. 131–142. Springer US. Riley, G. F., Sharif, M. I. and Lee, W. (2004) Simulating Internet worms. In Pro- ceedings of the 12th International Workshop on Modeling, Analysis, and Simu- lat

Early Stage Botnet Detection and Containment via ... - Semantic Scholar
this research is to localize weakly connected subgraphs within a graph that models network communications between .... ultimate objective behind such early state detection is botnet containment. The approach that we ...... 71–86. Staniford, S., Pax

Target Detection and Verification via Airborne ...
probability of detection, in comparison to detection via one of the steps only. ... He is now with the School of Computer Science and Engineering, The. Hebrew University .... We collected data at three different times of year: summer, spring, and ...

Background Calibration of Pipelined ADCs Via ...
Science and Engineering Graduate Fellowship (NDSEG), the MIT Center for. Integrated Circuits and Systems, and ... Science, Massachusetts Institute of Technology, Cambridge, MA 02139 USA. (e-mail: [email protected]). ...... [19] U. Eduri and F. Malobert

Anomaly Detection via Optimal Symbolic Observation of ...
Condition monitoring and anomaly detection are essential for the prevention of cascading ..... no false alarm, and no missed detection of special events (such as ...

Real-Time Detection of Malware Downloads via - UGA Institute for ...
Mar 19, 2016 - toriously ineffective against malware code obfuscation [11], whereas URL blacklists can often be .... clients who voluntarily agree to share information about file download events. In addition, the data is .... if u shares the same URL

Hierarchical Co-salient Object Detection via Color Names - GitHub
HCN. Figure 2: Illustration of single-layer combination. Border Effect. In the testing data set, some of the input images have thin artificial borders, which may affect.

Early Detection of Malicious Flux Networks via Large ...
Index Terms—Flux Networks, DNS, Passive Traffic Analysis, Clustering, Classification, Internet Security .... Because we use a different type of data source than the one we .... means, including, for example, blog spam, social websites spam ...

Salient Region Detection via High-Dimensional Color Transform
As demonstrated in our experimental results, our per-pixel saliency map represents how distinctive the color of salient regions is compared to the color of the background. Note that a simple linear combination or transformation of the color space can

Eye Localization via Eye Blinking Detection
May 1, 2006 - Jerry Chi-Ling Lam (994873428). Department of Electrical and Computer Engineering, University of Toronto, 10 Kings College Road. Toronto ...

Signal detection via residence-time asymmetry in noisy ...
Jan 31, 2003 - quite large, often above the deterministic switching threshold that is itself ...... issue is the amount of data dependent on the response time.

Eye Localization via Eye Blinking Detection
In this paper, the goal is to localize eyes by determining the eye blinking motions using optical flow happened in a sequence of images. Frame differencing is used to find all possible motion regions that satisfy the eye blinking mo- tion constraints

Anomaly Detection via Online Over-Sampling Principal Component ...
... detection has been an important research topic in data mining and machine learning. Many real- ... detection in cyber-security, fault detection, or malignant ..... Anomaly Detection via Online Over-Sampling Principal Component Analysis.pdf.

Signal detection via residence-time asymmetry in noisy ...
Jan 31, 2003 - 2 / ), since an analytic solution of. Eq. 1 is not possible for large bias signal amplitudes. We note that in practical devices, the bias signal is ...