Image Annotations Based on Semi-supervised Clustering with Semantic Soft Constraints Rui Xiaoguang, Yuan Pingbo, and Yu Nenghai* MOE-Microsoft Key Laboratory of Multimedia Computing and Communication University of Science and Technology of China Hefei, Anhui, China [email protected], {ypb, ynh}@ustc.edu.cn

Abstract. An efficient image annotation and retrieval system is highly desired for the increase of amounts of image information. Clustering algorithms make it possible to represent images with finite symbols. Based on this, many statistical models, which analyze correspondence between visual features and words, have been published for image annotation. But most of these models cluster only using visual features, ignoring semantics of images. In this paper, we propose a novel model based on semi-supervised clustering with semantic soft constraints which can utilize both visual features and semantic meanings. Our method first measures the semantic distance with generic knowledge (e.g. WordNet) between regions of the training images with manual annotations. Then a semisupervised clustering algorithm with semantic soft constraints is proposed to cluster regions with semantic soft constraints which are formed by semantic distance. The experiment results show that our model improves performance of image annotation and retrieval system. Keywords: image annotation, semi-supervised clustering, soft constraints, semantic distance.

1 Introduction With the rapid development of digital photography, digital image data has increased tremendously in recent years. Consequently image retrieval has drawn the attention of many researchers. Content-based image retrieval (CBIR) computes relevance based on the visual similarity of low-level image features. However there is a gap between low-level visual features and semantic meanings. The so-called semantic gap is a major problem that needs to be solved for most CBIR approaches. Consequently, image annotation which can settle this problem has received extensive attention recently. One approach to automatically annotate images is to look at the probability of associating words with image regions. Mori et al. [1] proposed a co-occurrence model in which they looked at the co-occurrence of words with image regions created using a regular grid. Duygulu et al [2] proposed to describe images using a vocabulary of *

Corresponding author.

Y. Zhuang et al. (Eds.): PCM 2006, LNCS 4261, pp. 624 – 632, 2006. © Springer-Verlag Berlin Heidelberg 2006

Image Annotations Based on Semi-supervised Clustering

625

blobs. First, regions are created using a segmentation algorithm like normalized cuts. For each region, features are computed and then blobs are generated by clustering the image features for these regions across images. Each image is generated by using a certain number of these blobs. Their Translation Model (TM) applies one of the classical statistical machine translation models to translate from the set of keywords of an image to the set of blobs forming the image. Jeon et al [3]used a cross-media relevance model (CMRM) to perform both image annotation and ranked retrieval. Since most of above approaches rely on clustering as the basis for automatic image annotation, the performance of annotation is strongly influenced by the quality of clustering. However, most approaches perform region clustering only based on lowlevel visual features, ignoring semantic concepts of images. Thus regions with different semantic concepts but share similar appearance may be grouped, leading to a poor clustering performance. To address this problem, we first measure the correlations of annotations based on semantic distance. Next, we select some of the semantic distance of annotations as soft constraints, and then develop a semi-supervised clustering with these semantic soft constraints when clustering the regions to blobs. So the new approach comprehensively uses both semantic concepts and low-level features. In the previous research some efforts have been made to cluster using a semisupervised method. The prior knowledge was provided at the instance level in the form of positive (must-link) and negative (cannot-link) pairwise constraints in [4-5]. Soft constraints were introduced in the dissertation of Wagsta [6]. [7] proposed a principled probabilistic framework based on Hidden Markov Random Fields (HMRFs) for semi-supervised clustering. Our clustering approach can be seen as an extend work by [7] whose method gets a good performance, but only copes with the hard constraints. To account for soft constraints, we use a more complex objective function. The main contribution of this paper is as follows: we propose an image annotation model based on semi-supervised clustering which can make use of the semantic meanings of images, and develop an image annotation and retrieval system. This paper is organized as follows: Section 2 explains semantic distance measure and how it forms the soft constraints, and presents our proposed approach of the semisupervise clustering which can utilize these semantic soft constraints. Section 3 presents experiment setup and results of our approach. Section 4 presents conclusion and a comment on future work.

2 Region Clustering with Semantic Soft Constraints Most approaches perform image regions clustering merely based on visual features. Thus regions with different semantic concepts but similar appearance may be easily grouped, which will lead to poor clustering performance. In TM & CMRM model, each training image is represented by a set of keywords and visual tokens, and every visual token inherits all the concepts from its image.

626

R. Xiaoguang, Y. Pingbo, and Y. Nenghai

That is to say, we can get the semantic meanings of training images as prior knowledge in an image annotation task. So by using the structure and content of WordNet or any other thesaurus, similarities between tokens can be calculated not only using visual token feature but also semantic concepts. A natural way to use these semantic similarities is to impose constraints on the processing of clustering. Then we develop a semi-supervised clustering approach with semantic soft constraints which consist of soft must-link / cannot-link constraints. So, the task contains the following parts: • Measure semantic distance between visual tokens. • Form soft constraints • Develop a semi-supervised clustering using these constraints 2.1 Measuring Semantic Distance We will use the structure and content of WordNet for measuring semantic similarity between two concepts. Current state of the art can be classified to different categories such as: Node-Based [8-10], Distance-Based [11]. In our system, we choose LIN measure. Lin et al. [9] used the first Information Content (IC) notion and took into account the similarity between selected concepts. Lin used Corpus to get the probabilities of each concept and computed how many times the concept appears in the Corpus. Next, the probabilities of each concept were calculated by the relative frequency. So the Information Content (IC) is determined. With regard to semantic similarity between two concepts, Lin used the IC values of these concepts along with the IC value of lowest common subsumer (lcs) of these two concepts. similarity (c1 , c2 ) =

2 × IC (lcs (c1 , c2 )) IC (c1 ) + IC (c2 )

(1)

Furthermore, we can get semantic distance between visual tokens: −1 t1.image = t2 .image ⎧⎪ D (t1 ,t 2 ) = ⎨1 − average[ min similarity (t .keyword , t .keyword )] t .image ≠ t .image p 2 q 1 1 2 ⎪⎩ min( p , q ) max( p , q )

(2)

Where t1 , t2 refer to visual tokens, t1 .image means the image which t1 belongs to,

t1 .keyword p means the p-th keyword with which t1 .image is manually annotated. When t1 , t2 belong to the same image, it is no use to calculate the semantic distance between them for their keywords are the same, so we set its value -1. 2.2 Form Soft Constraints

By using semantic distances between visual tokens, constraints can be formed. There are two common constraints: must-link constraints and cannot-link constraints.

Image Annotations Based on Semi-supervised Clustering

627

A cannot-link/ must-link constraint means two data (in our work, it is a token) cannot/ must be put in the same cluster. They are always “hard” constraints, while in our work we can define a constraint with an additional strength factor which is called a “soft” constraint. Compared with a hard constraint [12], a soft constraint has many advantages in our task. For example, there are two certain tokens: token A in the image“sky trees” and token B in another image“sky grass’. It is hard to use hard must-link constraints to say that token A and token B must in the same cluster, but it can be described by a soft must-link constraint with a strength factor s. If the semantic distance D between token A and B is smaller than the down-threshold, we think they have a soft mustlink constraint with a strength factor s = D. So, a soft constraint can better utilize known knowledge than a hard constraint, and can easily cope with semantic distance. Constraints can be listed using the following expression: if up - threshold < D < 1 ⎧ -D S= ⎨ ⎩1 − D if 0 < D < down - threshold

(3)

( A, B, S ) : define a constraint; A,B: tokens; S :strength factor

If semantic distance is out of threshold, a constraint will be formed. When S>0, (A, B, S) defines a soft must-link constraint; When S<0, (A, B, S) defines a soft cannotlink constraint. 2.3 Clustering with Soft Constraints

After forming soft constraints between regions from different images, we perform clustering with these constraints to generate region clusters/blobs. 2.3.1 HMRF-Kmeans HMRF-Kmeans [7] motivates an objective function derived from the posterior energy of the HMRF framework. J obj =

∑ D( x , μ i

xi ∈X

+



( xi , x j )∈C

li

)+



( xi , x j )∈M

wij D( xi , x j ) I (li ≠ l j )

wij ( Dmax − D( xi , x j )) I (li = l j )

(4)

Where M is the set of must-link constraints, C is the set of cannot-link constraints, wij / wij is the penalty cost for violating a must-link/cannot-link constraint between xi and xj, and li refers to the cluster label of xi. μli refers to the representative of cluster li . I is an indicator function (I[true] = 1, I[false] = 0). The first term in this objective function is the standard Kmeans’ objective function; the second term is a penalty function for violating must-link constraints; while the third term is a penalty function for violating cannot-link constraints. The algorithms developed to find a minimal value of this objective function, using an iterative relocation approach like Kmeans.

628

R. Xiaoguang, Y. Pingbo, and Y. Nenghai

2.3.2 Soft HMRF-Kmeans But HMRF-Kmeans can only cope with hard constraints. We achieve clustering with soft constraints by modifying HMRF-Kmeans’ objective function to deal with a realvalued penalty for violating constraints. Two methods are proposed: • Indicator method In the original objective function, Indicator function I (a) is a two-value function.

( ( I (true) = 1, I ( false) = 0) ) In order to deal with soft constraints, Indicator function can be modified to a real function I ′(a) range from 0 to 1. And the value of real indicator function I ′(a) equals the absolute value of the strength factor S when a is true. ( ( I (true) = S , I ( false) = 0) ) So in the modified objective function, the penalty is

proportional to the absolute value of the strength factor S. • MVS method MVS ( xi , μ h ) means in all soft constraints ( xi ,B,S) if xi in cluster h calculating the maximum strength of the violated constraints. MVS ( xi , μ h ) can be calculated as follows:

MVS ( xi , μ h ) =

nViol × max S nConst violate constraints

(5)

Where nConst is the times when S gets maximum value, nViol is the times when S gets maximum value and constraint is violated. New objective function combines MVS with old objective function in the following way: J oMbjS ( x i , μ h ) =

J obj ( x i , μ h ) 1 − M V S ( xi , μ h )

(6)

MS will reach a very high value. In When MVS gets a value which is near to 1, J obj

this case, xi will not choose cluster h. Therefore, the new objective function can show how strong constraints are. Finally, we combine these two methods, and get the soft objective function: SO FT J obj ( xi , μ h ) = [ D ( x i , μ h ) + ∑ wij D ( x i , x j ) I ′ ( li ≠ l j ) S >0

-∑

S <0

wij D ( x i , x j ) I ′( li = l j )] /(1 − M VS ( xi , μ h ))

(7)

Note that we remove Dmax for accelerating the processing of the algorithm and it will not alter the meaning of objective function. Soft HMRF-Kmeans aims to minimize the penalty of violating the soft constraints, which can be achieved by minimizing the new objective function. In detail, Soft HMRF-Kmeans is also an EM-like algorithm:

Image Annotations Based on Semi-supervised Clustering

629

− In the E-step, given the current cluster representatives, every data point is reSOFT assigned to the cluster which minimizes its contribution to J obj . Iterated SOFT conditional mode (ICM) [13] is applied to reduce J obj .

− In the M-step, the cluster representatives {μn }hK=1 are re-estimated from the cluster SOFT assignments to minimize J obj for the current assignment. Also the clustering distance measure D is updated to reduce the objective function by transforming the space. We define L2 distance measure D = A*|X-Y|, every weight am in A would be updated using the update rule: am = am + μ

SOFT ∂J obj

am

3 Experiment and Result 3.1 Dataset and Model Parameter

We use the dataset in Duygulu et al[2]. The dataset consists of 5,000 images from 50 Corel Stock Photo cds. Each cd includes 100 images on the same topic. Segmentation using normalized cuts followed by quantization ensures that there are 1-10 blobs for each image. Each image was also assigned 1-5 keywords. Overall there are 374 words in the dataset. Dataset is divided into 2 parts - the training set of 4500 images and the test set of 500 images. Two thresholds need to be defined to form constraints, and we choose up-threshold = 0.9, down-threshold = 0.3 which are tuned according to the experiment. Unit constraint costs W and W are used for all constraints, since indicator function already provides individual weights for the constraints. Other parameter such as constraints can be automatically handled by our model. 3.2 Results and Discussion

We perform image annotations based on CMRM model [3] and compare Soft HMRFKmeans/CMRM with Hard HMRF-Kmeans/CMRM as well as Kmeans/CMRM. • Soft HMRF-Kmeans/CMRM: use Soft HMRF-Kmeans in CMRM model. • Hard HMRF-Kmeans/CMRM: use Hard HMRF-Kmeans in CMRM model. And it only uses cannot-link constraints because it is difficulty to form hard must-link constraints according to the Section 2.2. • Kmeans/CMRM: only use unsupervised Kmeans in CMRM model.

Fig. 1 shows some annotation results based on these three models. Fig. 2 is a precise-recall graph and demonstrates the performance of the retrieval system. Comparing these three image annotation models, system performance enhances when using a semi-supervised clustering method. And soft HMRF-Kmeans/CMRM gets the best performance.

630

R. Xiaoguang, Y. Pingbo, and Y. Nenghai

Image

Original annotation

boats horizon shops water water sky man people clouds water sky clouds pillar buildings boats ships water sky vehicle

Kmeans/CMRM Hard HMRFKmeans/CMRM Soft HMRFKmeans/CMRM

mountain people road tree temple buddha people sky tree people buildings street swimmers people street road grass mountain

beach people sand water sky sand hills dunes water sand sky dunes people water sand beach dunes people water

Fig. 1. Automatic annotations (best five words) compared with the original manual annotations

Recall/Precision Graph 0.4

Soft HMRF-Kmeans/CMRM Kmeans/CMRM Hard HMRF-Kmeans/CMRM

0.35 0.3

Precision

0.25 0.2

0.15 0.1 0.05 0 0

0.2

0.4

0.6

0.8

1

Recall

Fig. 2. Performance of the three model

We can see more detail in Fig. 1. The automatic annotations of Kmeans/CMRM often contain some irrelevant keywords, e.g. “man”, “people” in first image. And the semi-clustering/CMRM can remove some of them, e.g. “people” (first image), “sky” (second image), “hills” (third image). Comparing with Hard HMRF-Kmeans/CMRM, Soft HMRF-Kmeans/CMRM can remove more irrelevant keywords (e.g. “sky” in third image) and extend the existed keywords (e.g. “boats” in the first image is extended with “ships” and “vehicle”; “street” which is the automatic annotations of Hard HMRF-Kmeans/CMRM in the second image is extended with “road”, and the same is to “sand” in the third image). Therefore, when apply Soft HMRF-Kmeans to cluster regions to blobs, our system naturally have the ability of “removing” and “extending” when using semantic constraints. Manual annotations are not always true, e.g. “shops” (first image). But In the Soft HMRF-Kmeans/CMRM model, it is corrected with “ships”. So this may be a useful approach in checking the accuracy of manual annotations.

Image Annotations Based on Semi-supervised Clustering

631

Fig.3 shows that blobs formed by Kmeans are not well described. The test image and the training image are 5 blobs in common while they have not any semantic relation. But blobs formed by Soft HMRF-Kmeans are described better. The two images have little blobs in common. So blobs formed by Soft HMRF-Kmeans are better descriptors of images.

Left: training image Right: test image Soft HMRF-Kmeans Kmeans

Blobs: 56 95 259 458 62 56 106 256 Blobs: 83 419 186 8 33 33 250 142 142 199

Blobs: 362 507 159 20 64 92 106 Blobs: 449 145 8 484 33 142 145 142 269

Fig. 3. different clustering method to form blobs

4 Conclusions and Future Works We have shown a semi-supervised clustering can help existed region-based image annotation model improve the performance of annotating and retrieving image. Using semantic meanings of regions, soft constraints can be formed to get better results. Obtaining large amounts of labeled training and test data is difficult but we believe this is needed for improvements in both performance and evaluation of the algorithms proposed here. Better feature extraction will probably improve the results. Other areas of possible research include the use of captions in the World Wide Web. We believe that this is a fruitful area of research for applying semi-supervised learning method in image or video annotation.

Acknowledgement This work is supported by MOE-Microsoft Key Laboratory of Multimedia Computing and Communication Open Foundation (No.05071804). The authors wish to thank Prof. Li Mingjing for his contributions in this work.

References 1. Mori Y, Takahashi H, Oka R. Image-to-word transformation based on dividing and vector quantizing images with words. First International Workshop on Multimedia Intelligent Storage and Retrieval Management, 1999 2. Duygulu P, Barnard K, De F N, et al. Object recognition as machine translation: Learning a lexicon for a fixed image vocabulary. Seventh European Conference on Computer Vision (ECCV), 2002, 4: 97 112 3. Jeon J, Lavrenko V, Manmatha R. Automatic image annotation and retrieval using crossmedia relevance models. Toronto, Canada: ACM Press, 2003. 119 126P





632

R. Xiaoguang, Y. Pingbo, and Y. Nenghai

4. Wagstaff K, Cardie C, Rogers S, et al. Constrained k-means clustering with background knowledge. Proceedings of the Eighteenth International Conference on Machine Learning, 2001, 577 584 5. Wagstaff K, Cardie C. Clustering with instance-level constraints. Proceedings of the Seventeenth International Conference on Machine Learning, 2000, 1103 1110 6. Wagstaff K et al. Intelligent Clustering with Instance-Level Constraints, Proceedings of the Seventeenth International Conference on Machine Learning, 2000, 1103 7. Basu S, Bilenko M, Mooney R J. A probabilistic framework for semi-supervised clustering. Proceedings of the 2004 ACM SIGKDD international conference on Knowledge discovery and data mining, 2004, 59 68 8. Jiang J J, Conrath D W. Semantic similarity based on corpus statistics and lexical taxonomy. Proceedings of International Conference on Research in Computational Linguistics, 1997, 19 33 9. Lin D. Using syntactic dependency as local context to resolve word sense ambiguity. Proceedings of the 35th Annual Meeting of the Association for Computational Linguistics, 1997, 64 71 10. Resnik P. Using information content to evaluate semantic similarity in a taxonomy. Proceedings of the 14th International Joint Conference on Artificial Intelligence, 1995, 1: 448 453 11. Leacock C, Chodorow M. Combining local context and WordNet similarity for word sense identification. WordNet: An Electronic Lexical Database, 1998, 265 283 12. Shi R, Wanjun J, Tat-seng C. A Novel Approach to Auto Image Annotation Based on Pairwise Constrained Clustering and Semi-Naive Bayesian Model. 2005. 322 327P 13. Besag J. On the statistical analysis of dirty pictures (with discussion). Journal of the Royal Statistical Society, Series B, 1986, 48(3): 259 302





~1110.















LNCS 4261 - Image Annotations Based on Semi ... - Springer Link

Keywords: image annotation, semi-supervised clustering, soft constraints, semantic distance. 1 Introduction ..... Toronto, Canada: ACM Press, 2003. 119~126P ...

378KB Sizes 1 Downloads 264 Views

Recommend Documents

LNCS 4261 - Image Annotations Based on Semi ... - Springer Link
MOE-Microsoft Key Laboratory of Multimedia Computing and Communication ..... of possible research include the use of captions in the World Wide Web. ... the Seventeenth International Conference on Machine Learning, 2000, 1103~1110.

LNCS 4191 - Registration of Microscopic Iris Image ... - Springer Link
Casey Eye Institute, Oregon Health and Science University, USA. {xubosong .... sity variance in element m, and I is the identity matrix. This is equivalent to.

LNCS 6621 - GP-Based Electricity Price Forecasting - Springer Link
learning set used in [14,9] at the beginning of the simulation and then we leave ..... 1 hour on a single core notebook (2 GHz), with 2GB RAM; the variable ...

LNCS 7335 - Usage Pattern-Based Prefetching: Quick ... - Springer Link
Oct 8, 2010 - The proposed scheme is implemented on both Android 2.2 and Linux kernel 2.6.29. ... solution for reducing page faults [1, 2, 3, 10]. The number ...

LNCS 6621 - GP-Based Electricity Price Forecasting - Springer Link
real-world dataset by means of a number of different methods, each cal- .... one, that we call GP-baseline, in which the terminal set consists of the same variables ...

LNCS 6622 - NILS: A Neutrality-Based Iterated Local ... - Springer Link
a new configuration that yields the best possible fitness value. Given that the .... The neutral degree of a given solution is the number of neutral solutions in its ...

LNCS 6683 - On the Neutrality of Flowshop Scheduling ... - Springer Link
Scheduling problems form one of the most important class of combinatorial op- .... Illustration of the insertion neighborhood operator for the FSP. The job located.

LNCS 3352 - On Session Identifiers in Provably Secure ... - Springer Link
a responding server process in network service protocol architecture [23]. A ... 3. proposal of an improved 3PKD protocol with a proof of security using a.

LNCS 4731 - On the Power of Impersonation Attacks - Springer Link
security or cryptography, in particular for peep-to-peer and sensor networks [4,5]. ... entity capable of injecting messages with arbitrary content into the network.

LNCS 6361 - Automatic Segmentation and ... - Springer Link
School of Eng. and Computer Science, Hebrew University of Jerusalem, Israel. 2 ... OPG boundary surface distance error of 0.73mm and mean volume over- ... components classification methods are based on learning the grey-level range.

LNCS 6942 - On the Configuration-LP for Scheduling ... - Springer Link
insights on two key weaknesses of the configuration-LP. For the objective of maximizing the minimum machine load in the unrelated graph balancing setting ...... length. European Journal of Operational Research 156, 261–266 (2004). 19. Scheithauer,

LNCS 6719 - Multiple People Activity Recognition ... - Springer Link
Keywords: Multiple Hypothesis Tracking, Dynamic Bayesian Network, .... shared space and doing some basic activities such as answering phone, using.

LNCS 3174 - Multi-stage Neural Networks for Channel ... - Springer Link
H.-S. Lee, D.-W. Lee, and J. Lee. In this paper, we propose a novel multi-stage algorithm to find a conflict-free frequency assignment with the minimum number of total frequencies. In the first stage, a good initial assignment is found by using a so-

LNCS 4325 - An Integrated Self-deployment and ... - Springer Link
The VFSD is run only by NR-nodes at the beginning of the iteration. Through the VFSD ..... This mutual effect leads to Ni's unpredictable migration itinerary. Node Ni stops moving ... An illustration of how the ZONER works. The execution of the ...

LNCS 650 - Emergence of Complexity in Financial ... - Springer Link
We start with the network of boards and directors, a complex network in finance which is also a social ... from the board of Chase Manhattan Bank. Boards of ...

LNCS 4233 - Fast Learning for Statistical Face Detection - Springer Link
Department of Computer Science and Engineering, Shanghai Jiao Tong University,. 1954 Hua Shan Road, Shanghai ... SNoW (sparse network of winnows) face detection system by Yang et al. [20] is a sparse network of linear ..... International Journal of C

LNCS 3973 - Local Volatility Function Approximation ... - Springer Link
S&P 500 call option market data to illustrate a local volatility surface estimated ... One practical solution for the volatility smile is the constant implied volatility approach .... Eq. (4) into Eq. (2) makes us to rewrite ˆσRBF (w; K, T) as ˆσ

LNCS 7601 - Optimal Medial Surface Generation for ... - Springer Link
parenchyma of organs, and their internal vascular system, powerful sources of ... but the ridges of the distance map have show superior power to identify medial.

LNCS 3715 - Security of Two-Party Identity-Based Key ... - Springer Link
technology, namely a saving in computation and key size. This is ...... Conference on Networking Technologies, Services, and Protocols - NETWORKING. 2004 ...

LNCS 3715 - Security of Two-Party Identity-Based Key ... - Springer Link
K. Y. Choi, J. Y. Hwang, D. H. Lee, and I. S. Seo. ID-based Authenticated Key. Agreement for Low-Power Mobile Devices. In 10th Australasian Conference on.

LNCS 4258 - Privacy for Public Transportation - Springer Link
Public transportation ticketing systems must be able to handle large volumes ... achieved in which systems may be designed to permit gathering of useful business ... higher powered embedded computing devices (HPDs), such as cell phones or ... embedde