Pattern Recognition Letters 24 (2003) 2255–2267 www.elsevier.com/locate/patrec

Finding textures by textual descriptions, visual examples, and relevance feedbacks q Hsin-Chih Lin a

a,*

, Chih-Yi Chiu b, Shi-Nine Yang

b

Department of Information Management, Chang Jung Christian University, 396 Chang Jung RD., Sec. 1, Tainan County 711, Taiwan b Department of Computer Science, National Tsing Hua University, Hsinchu 300, Taiwan Received 28 March 2002; received in revised form 2 January 2003

Abstract In this study, we propose a fuzzy logic CBIR (content-based image retrieval) system for finding textures. In our CBIR system, a user can submit textual descriptions and/or visual examples to find the desired textures. After the initial search, the user can give relevant and/or irrelevant examples to refine the query and improve the retrieval efficiency. Contributions of this study are fourfold. (1) Our CBIR system maps low-level statistical features to high-level textual concepts; it bridges the semantic gap between these two levels. (2) Our CBIR system characterizes texture properties of these two levels; and further, it achieves high-level texture manipulations through textual concepts. (3) Our CBIR system models the human perception subjectivity via relevance feedbacks to perform more accurate retrieval. (4) Our CBIR system provides intuitive and simple methods of similarity definitions and computations. Experimental results reveal our CBIR system is indeed effective. The retrieved images are perceptually satisfactory, and the retrieval time is very short. Ó 2003 Elsevier B.V. All rights reserved. Keywords: Fuzzy logic system; Content-based image retrieval (CBIR); Texture retrieval; Relevance feedback; Semantic gap; Human perception subjectivity

1. Introduction Content-based image retrieval (CBIR) is an emerging research topic for multimedia databases

q

This study was supported partially by National Science Council, R.O.C. under Grant NSC89-2218-E-309-001 and Ministry of Education, R.O.C. under Grant 89-E-FA04-1-4. * Corresponding author. E-mail addresses: [email protected] (H.-C. Lin), [email protected] (C.-Y. Chiu), [email protected] (S.-N. Yang).

and digital libraries. Since the number of images grows rapidly in todayÕs digital archives and computer networks, effective techniques for finding images in a large repertory are urgently required. By content-based techniques, a user can query an image database by contents of interest, which may be colors, textures, shapes, and the spatial layout of target images. Images that satisfy ‘‘perceptual similarity’’ to the userÕs query can be found in the database. Many CBIR systems have been proposed during the last decade. Commercial systems include QBIC, Virage, RetrievalWare, and

0167-8655/03/$ - see front matter Ó 2003 Elsevier B.V. All rights reserved. doi:10.1016/S0167-8655(03)00052-7

2256

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

so on. Research systems include Photobook, WebSEEk, Netra, MARS, and so on. Detailed surveys on CBIR systems can be referred to (Del Bimbo, 1999; Aigrain et al., 1996; Chang et al., 1997; Gupta and Jain, 1997; Idris and Panchanathan, 1997; Rui et al., 1999; Smeulders et al., 2000; Goodrum, 2000; Shapiro and Stockman, 2001). In our earlier work (Lin et al., 1997), a CBIR system was proposed so that a user can specify image regions and their spatial layout in a query. Images that partially or exactly match the query can be effectively retrieved. Unfortunately, if the desired images have texture-like patterns, retrieval results may be unsatisfactory. The world is rich in textures. No formal definition of textures exists despite their importance and ubiquity in digital imaging. The definition of textures varies in each application. Although texture analysis has been studied for several decades, research results obtained to date are by no means perfect. For that reason, many researchers still keep on contributing texture analysis techniques. In-depth studies on texture analysis can be referred to (Haralick, 1982; Nadler and Smith, 1993; Tuceryan and Jain, 1998) studies related to our work are discussed as follows. Haralick et al. (1973) proposed co-occurrence matrix (CM) representations to solve the classification problem of texture images. These representations describe spatial dependence of pixels over a gray-scale image. Fourteen statistical features derived from co-occurrence matrices are used to describe various texture properties. Among the 14 CM-features, Conners and Harlow (1980) reported that only five were truly useful in real applications. The five CM-features include energy, entropy, correlation, local homogeneity (or inverse different moment), and inertia. Afterwards, Lin et al. (1999) proposed a CBIR system based on the five CM-features. In LinÕs system, a regulartexture image is represented as one texture primitive and five CM-features. The texture primitive is effective in describing periodic properties over a regular-texture image; the CM-features are satisfactory for describing statistical properties within the primitive. As a result, the comparison of two images can be achieved by simple comparisons of texture primitives and CM-features. Note

that LinÕs system deals with regular textures only. Tamura et al. (1978) proposed a texture representation based on psychological studies of human perceptions. This representation consists of six statistical features, including coarseness, contrast, directionality, line-likeness, regularity, and roughness, to describe various texture properties. Tamura features are visually meaningful, whereas some of CM-features (e.g., entropy and inertia) may not. This advantage makes Tamura features very attractive in texture-based image retrieval. For example, Flickner et al. (Niblack and Flickner, 1993; Flickner et al., 1995; Niblack et al., 1993; Ashley et al., 1995) used three of Tamura features, including coarseness, contrast, and directionality, to design the QBIC system, in which a texture image is represented as the three features. The image comparison is achieved by evaluating the weighted Euclidean distance in the 3D feature space; each weight is the inverse variance on the feature. Rui et al. (1999) also used the three features in the MARS system. Liu and Picard (1996) presented a Wold representation for textures in the Photobook system (Pentland et al., 1994), in which a texture image is regarded as a homogeneous 2D discrete random field, which can be decomposed into three features, including periodicity, directionality, and randomness. The Wold representation can effectively model the human perception subjectivity for finding similar textures; it produces compact texture descriptions but preserves perceptual attributes of textures. The approaches for texture analysis can be categorized into three classes: including statistical, structural, and spectral approaches. No matter what approach is used, a texture is usually represented as a low-level numerical vector, holding measures on certain texture properties. Texture analysis through numerical vectors may suffer the following limitations (Smeulders et al., 2000; Rui et al., 1998): Semantic gap between low-level features (e.g., numerical vectors) and high-level concepts (e.g., textual descriptions): The semantic gap comes from the inconsistency between features extracted from a texture image and user interpretations for the same image. The gap exists between the two semantic levels.

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

Texture Images

Texture Analysis

Similarity Inference

Tamura Features Fuzzy Clustering

Term Sets

user may also have difficulties in refining the query on certain texture features. To overcome the above drawbacks, we propose a fuzzy logic CBIR system, named as LinStar Texture (i.e., Linguistic Star for Textures), for finding textures through textual descriptions, visual examples, and relevance feedbacks. LinStar is an on-going project whose goal is to incorporate high-level textual concepts into multimedia retrieval systems. A user can easily submit a right query through the graphical user interface. After the initial search, the user can give relevant and/or irrelevant examples to refine the query for more accurate retrieval. Our system architecture consists of two major phases, as shown in Fig. 1, including the database creation and query comparison phases; its overview is given as follows. Texture analysis: A texture image is represented as six Tamura features, which characterize lowlevel statistical properties of textures. Fuzzy clustering: A term set on each Tamura feature is generated through our proposed fuzzy clustering algorithm so that degrees of appearance for the feature can be interpreted as our proposed five linguistic terms, which characterize high-level textual concepts of textures.

Linguistic Terms

On Line

Off Line

Human perception subjectivity: Different users, or the same user under different circumstances, may perceive the same texture differently. Moreover, the way users define the similarity between two textures may be quite different. The subjectivity exists in each semantic level. To improve the effectiveness of texture analysis, incorporating high-level concepts, such as textual descriptions, may be promising. Textual descriptions are the most natural way of human communications; they also yield more informative and reliable interpretations for texture images. If we can map low-level features to high-level concepts, the semantic gap between the two semantic levels may be bridged. Unfortunately, the mapping is by no means easy; it depends on the user perception subjectivity and cultural backgrounds. Meanwhile, the study on high-level texture analysis is still in its infancy. For that reason, most of existing CBIR systems provide methods of querying by examples (QBEs), in which texture features are transparent to users. Instead of specifying low-level texture features, a user can give a visual example as a query to find the desired textures. The QBE method is an effective query scheme. However, the user may be hard to find a right example from a large image database. After the initial search, the

Tamura Features

Query Parsing

Texture Analysis

Textual Descriptions

Visual Examples

Similarity Functions Similarity Computation

Sorted Indices

Similar Images

Image Display

GUI Image Database

(a)

2257

(b) Fig. 1. System architecture: (a) database creation; (b) query comparison.

2258

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

Image database: Each texture image, together with its Tamura features and linguistic terms, is organized and stored in the image database. The generated term sets are also stored in the database. Similarity definition: A query is expressed as a logic composition of linguistic terms or (Tamura) feature values. Therefore, the similarity between the query and a texture image can be interpreted as the similarity between the query expression and the image representation. We propose methods of defining similarity functions for linguistic terms or feature values. Similarity computation: We propose methods of computing the similarity between the query expression and the image representation through min–max composition rules. The rest of this paper is organized as follows. Section 2 describes the database creation phase. Section 3 describes the query comparison phase. Section 4 reports on experimental results. Conclusions and further research are given in the last section.

2. Database creation The database creation phase deals with texture analysis and fuzzy clustering. The goal of texture analysis is to extract statistical features from each texture image. The methods of texture analysis are described in Section 2.1. The goal of fuzzy clustering is to generate a term set on each feature so that we can map the low-level statistical feature to high-level textual concepts. The methods of fuzzy clustering are described in Section 2.2. 2.1. Texture analysis In our CBIR system, each texture image in the database is represented as six Tamura features. Although Tamura features characterize low-level statistical properties of textures, they have been shown visually meaningful and can be easily interpreted through high-level textual concepts. ThatÕs why Tamura features are used in our CBIR system. Detailed discussions about Tamura features can be referred to (Tamura et al., 1978); we summarize their visual properties and computa-

tions as follows (Del Bimbo, 1999; Haralick, 1982; Nadler and Smith, 1993; Tamura et al., 1978). Coarseness: Coarseness is the most fundamental feature in texture analysis; it refers to texture granularity, that is, the size and number of texture primitives. A coarse texture contains a small number of large primitives, whereas a fine texture contains a large number of small primitives. Coarseness (fcrs ) can be computed as follows. n X n 2k X pði; jÞ; fcrs ¼ 2 n i j where n  n denotes the image size, the sum is carried out for every pixel pði; jÞ and k is obtained as the value which maximizes P Pthe differences of the moving averages ð1=22k Þ pði; jÞ, taken over a 2k  2k neighborhood, along the horizontal and vertical directions. In this study, k ¼ 1, 2, 3, 4, and 5. Contrast: Contrast stands for image quality in the narrow sense; it refers the difference in intensity among neighboring pixels. A texture on high contrast has large difference in intensity among neighboring pixels, whereas a texture on low contrast has small difference. Contrast (fcon ) can be computed as follows. r ; fcon ¼ ðl4 =r4 Þ1=4 where r is the image standard deviation and l4 is fourth moment of the image. Directionality: Directionality is a global property over a specific region; it refers the shape of texture primitives and their placement rule. A directional texture has one or more recognizable orientation of primitives, whereas an isotropic texture has no recognizable orientation of primitives. Directionality (fdir ) can be computed as follows. fdir ¼ 1  r  np 

np X X p

2

ð/  /p Þ  HD ð/Þ;

/2wp

where HD is the local direction histogram, np is the number of peaks of HD , /p is the pth peak position of HD , wp is the range of pth peak between valleys, r is a normalizing factor, and / is the quantized direction code. In this study, the number of bins

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

for HD is 16, the quantized direction code / ¼ 0; 1; 2; . . . ; 15, and the normalizing factor r ¼ 0:025. Line-likeness: Line-likeness refers only the shape of texture primitives. A line-like texture has straight or wave-like primitives whose orientation may not be fixed. Often the line-like texture is simultaneously directional. Line-likeness (flin ) can be computed as follows.  , n X n X 2p flin ¼ PDd ði; jÞ cos ði  jÞ n i j n X n X i

PDd ði; jÞ;

j

where PDd ði; jÞ is the n  n local direction cooccurrence matrix of points at a distance d. Regularity: Regularity refers to variations of the texture-primitive placement. A regular texture is composed of identical or similar primitives, which are regularly or almost regularly arranged. An irregular texture is composed of various primitives, which are irregularly or randomly arranged (Haralick, 1982). Regularity (freg ) can be computed as follows. freg ¼ 1  rðrcrs þ rcon þ rdir þ rlin Þ; where r is a normalizing factor and rxxx means the standard deviation of fxxx . In this study, the normalizing factor r ¼ 0:25. Roughness: Roughness refers tactile variations of physical surface. A rough texture contains angular primitives, whereas a smooth texture contains rounded blurred primitives. Roughness (frgh ) can be computed as follows. frgh ¼ fcrs þ fcon :

2259

Tamura et al. (1978) concluded that coarseness, contrast, and directionality achieve successful correspondences with psychological measurements. However, line-likeness, regularity, and roughness require further improvement due to their discrepancies with psychological measurements. Although the use of the six Tamura features in our CBIR system is straightforward, experimental results show that Tamura features are effective in texture retrieval and interpretations. 2.2. Fuzzy clustering In fuzzy logic applications, choosing membership functions to reflect the data distribution is the first and an essential step; the choice may be achieved by using unsupervised or supervised learning algorithms (Zimmermann, 1991). In our CBIR system, a term set on each Tamura feature is generated through an unsupervised fuzzy clustering algorithm so that degrees of appearance for the feature can be interpreted as five linguistic terms. For example, linguistic terms for coarseness can be interpreted as ‘‘very fine,’’ ‘‘fine,’’ ‘‘medium coarse,’’ ‘‘coarse,’’ and ‘‘very coarse.’’ Linguistic terms for other Tamura features can be interpreted likewise, as summarized in Table 1. The degree of appearance increases from left to right in this table. The fuzzy clustering algorithm is presented as follows. Algorithm. Fuzzy Clustering Input. Data sequence x1 ; x2 ; . . . ; xn , where xi denotes a Tamura feature of the ith texture image and n is the number of texture images.

Table 1 Linguistic terms for the six Tamura features Tamura features

Linguistic terms

Coarseness Contrast Directionality Line-likeness Regularity Roughness

Very Very Very Very Very Very

fine low isotropic blob-like irregular smooth

Fine Low Isotropic Blob-like Irregular Smooth

Medium Medium Medium Medium Medium Medium

coarse on contrast directional line-like regular rough

Coarse High Directional Line-like Regular Rough

Very Very Very Very Very Very

coarse high directional line-like regular rough

2260

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

Output. Five membership functions of the Tamura feature, i.e., a term set on the Tamura feature. Step 1. Let c0 ¼ minfx1 ; x2 ; . . . ; xn g and c6 ¼ maxfx1 ; x2 ; . . . ; xn g. Compute c1 ; c2 ; . . . ; c5 as follows. j cj ¼ c0 þ  ðc6  c0 Þ: 6 Initialize membership functions as Fig. 2, in which c1 ; c2 ; . . . ; c5 denote class centers of the membership functions. Step 2. Set U ¼ 0. For each datum xj , update each element uij using one of the following rules, where uij , 1 6 i 6 5 and 1 6 j 6 n, is the membership value of xj in the ith linguistic term. Rule 1. If xi 6 c1 , ui;1 ¼ 1 and ui;k6¼1 ¼ 0. Rule 2. If cj < xi 6 cjþ1 , compute uij ¼ ðcjþ1  xi Þ=ðcjþ1  cj Þ, ui;jþ1 ¼ 1  uij , and ui;k6¼j;jþ1 ¼ 0. Rule 3. If xi > c5 , ui;k6¼5 ¼ 0 and ui;5 ¼ 1. Step 3. Compute c1 ; c2 ; . . . ; c5 using the following equation: Pn j¼1 uij xj c i ¼ Pn : ð1Þ j¼1 uij If c1 ; c2 ; . . . ; c5 are unchanged, the algorithm stops; otherwise go to Step 2. In the proposed algorithm, each linguistic term is a fuzzy set and represented as a triangular membership function. In Step 1, five evenly distributed triangular membership functions are used as initial membership functions of a Tamura feature, as shown in Fig. 2. c0 and c6 denote the

very low

low medium high

very high

u=1

u=0 c0

c1

c2

c3

c4

c5

c6

Tamura feature

Fig. 2. Membership functions of a Tamura feature.

minimum and maximum of the input data, respectively; c1 ; c2 ; . . . ; c5 denote class centers of the five membership functions. In Step 2, each element in U , i.e., uij , is updated according to the three rules. Rules 1–3 consider the cases as xj lies in the ranges ½c0 ; c1 , ðc1 ; c5 , and ðc5 ; c6 , respectively. In Step 3, the five class centers, i.e., c1 ; c2 ; . . . ; c5 are re-computed according to Eq. (1) to reflect new data distribution. If the five class centers are changed, the algorithm goes to Step 2 and starts a new iteration. If the five class centers are unchanged, the algorithm stops and the five membership functions are regarded as the representative ones. A representative term set is thus generated. In our CBIR system, Tamura features and linguistic terms characterize texture properties at various semantic levels. The generated term set formulates a mapping from low-level Tamura features to high-level linguistic terms; it also provides intuitive and simple methods of similarity definition and computations. Each texture image, together with its Tamura features and linguistic terms, is organized and stored in the image database. The generated term sets are also stored in the database. In the query comparison phase (Section 3), we will propose several methods of finding textures.

3. Query comparison In our CBIR system, a user can submit textual descriptions and/or visual examples to find the desired textures. After the initial search, the user can give relevant and/or irrelevant examples to refine the query and improve the retrieval efficiency. The method of querying by textual descriptions is described in Section 3.1. The method of querying by visual examples is described in Section 3.2. The method of relevance feedbacks is described in Section 3.3. 3.1. Querying by textual descriptions If the user submits a textual description to find the desired images, the query can be expressed as a

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267 I

Texture Image

R

Image Representation

ve ry fine

2261

fine medium coarse

ve ry coarse

u=1

fcrs f con f dir

f lin

f reg

f rgh

Tamura Features

u=0

coarseness c0

u crs u con u di r u lin u reg u rgh

tcrs t con t dir

t lin

t reg

t rgh

Similarity Functions

c2

c1

c3

c4

c5

c6

(a)

Final Similarity

min(max(u crs , u reg ), u co n )



Linguistic Terms

∧ max(u crs , u reg )

Query Expression

ve ry high on co ntrast





E

u crs Q

u co n fcon = very high

u reg

Textual Description

Fig. 3. Similarity model for a texture image and a textual description.

logic composition of linguistic terms. The logic operators include AND (denoted by ^), OR (_), and NOT (:). Therefore, similarity between the query and a texture image can be interpreted as the similarity between the query expression and the image representation. The similarity model is shown in Fig. 3. According to this model, the similarity function for each linguistic term (txxx ) must be defined. In addition, the method of aggregating multiple similarity functions to compute the final similarity must be proposed. In what follows, we will demonstrate through an example how the final similarity are computed. Consider a textual description ‘‘(fine _ regular) ^ very high on contrast’’ as a query. The similarity function for the linguistic term ‘‘fine’’ is defined as its membership function in the generated term set on coarseness, as shown in Fig. 4(a). Similarly functions for other linguistic terms are defined likewise. The query can be represented as a tree, as shown in Fig. 4(b), in which each leaf corresponds to a linguistic term and each internal node corresponds to a logic operator. After that, the final similarity can be computed by aggregating the three similarity functions through min–max composition rules, as shown in Fig. 4(c). In this example, the final similarity is minðmaxðucrs ; ureg Þ; ucon Þ.

fine

regu lar

f crs = fine

(b)

f reg = regular

(c)

Fig. 4. Querying by textual descriptions: (a) similarity definition for ‘‘fine;’’ (b) tree representation for the query; (c) similarity computations through min–max composition rules.

3.2. Querying by visual examples If the user submits a visual example with a specified constrain to find the desired images, the query can be expressed as a logic composition of (Tamura) feature values. Therefore, similarity between the query and a texture image can be interpreted as the similarity between the query expression and the image representation. The similarity model is shown in Fig. 5. According to this model, the similarity function for each feature value (gxxx ) must be defined. In addition, the method of aggregating multiple similarity functions to compute the final similarity must be proposed. In what follows, we will demonstrate through an example how the final similarity are computed. Consider a visual example with a specified constrain ‘‘similar on (coarseness _ regularity) ^ contrast’’ as a query. Let the coarseness value in the visual example be gcrs . Fig. 6(a) and (b) show similarity functions for gcrs in linguistic terms ‘‘medium coarse’’ and ‘‘coarse,’’ respectively. Adding the two partial similarity functions obtains the total similarity function for gcrs , as shown in

2262

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267 I

very fine

Texture Image

fine medium coarse

very coarse

u=1

R

fcrs f con f dir

Image Representation

f lin

f reg f rgh

u=0

Tamura Features

coarseness

c0

c2

c1

c 3 g crs c 4

c5

c6

(a) u crs u con u di r u lin u reg u rgh

Similarity Functions

Final Similarity

very fine

fine medium coarse

very coarse

u=1

g crs g con g di r g lin g reg g rgh

Tamura Features

E

Query Expression

u=0

coarseness

c0

c2

c1

c 3 g cr s c 4

c5

c6

(b) Visual Example

Fig. 6(c). Similarly functions for other feature values are defined likewise. The query can be represented as a tree, as shown in Fig. 6(d), in which each leaf corresponds to a feature value and each internal node corresponds to a logic operator. After that, the final similarity can be computed by aggregating the three similarity functions through min–max composition rules, as shown in Fig. 6(e). In this example, the final similarity is minðmaxðucrs ;ureg Þ; ucon Þ. 3.3. Querying by relevance feedbacks

very fine

fine medium coarse

coarseness

u=0 c0

c2

c1

c 3 g crs c 4

c5

c6

(c) min(max(u crs , u reg ), u con )



∧ max(u crs , u reg )

g co n u crs

g crs

g reg

(d)

After the initial search, the user can give relevant and/or irrelevant examples to refine the query and improve the retrieval efficiency. In what follows, we will demonstrate the method of giving relevant examples through an example. Assume a user gives two relevant examples with a specified constrain ‘‘find textures of similar coarseness, and so on.’’ Let the coarseness values in the two rele0 vant examples be gcrs and gcrs . Fig. 7(a) shows 0 similarity functions for gcrs and gcrs . If the user wishes to find images that are similar to both images, min-composing the two similarity functions obtains the similarity function used in the next search, as shown in Fig. 7(b). However, if the user wishes to find images that are similar to either of

very coarse

u=1



Fig. 5. Similarity model for a texture image and a visual example.



Q

f crs = g crs

u con f con = g con

u reg f reg = g reg

(e)

Fig. 6. Querying by visual examples: (a) partial similarity function for gcrs in ‘‘medium coarse;’’ (b) partial similarity function for gcrs in ‘‘coarse;’’ (c) similarity function for gcrs ; (d) tree representation for the query; (e) similarity computations through min–max composition rules.

the two images, max-composing the two similarity functions obtains the similarity function used in the next search, as shown in Fig. 7(c).

4. Experimental results LinStar Texture contains two texture image databases. The first database (i.e., our Corel data-

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267 very fine

fine medium coarse

very coarse

u=1

u=0

coarseness c0

c1

c2

c 3 g crs g' crs

c5

c6

(a) very fine

fine medium coarse

very coarse

u=1

u=0

coarseness c0

c1

c2

c 3 g crs g' crs

c5

c6

(b) very fine

fine medium coarse

very coarse

u=1

u=0

coarseness c0

c1

c2

c 3 g crs g' crs

c5

c6

(c) Fig. 7. Similarity function compositions: (a) two similarity functions; (b) min-composed similarity function (i.e., AND operation); (c) max-composed similarity function (i.e., OR operation).

base) contains 1570 192  128 texture images that are selected from Corel Gallery Collection. The second database (i.e., our VisTex database) is created as follows. First of all, we obtained 45 512  512 texture images from MIT VisTex. For each of the 45 images, 9 170  170 non-overlap sub-images were cropped into a group of relevant images. Consequently, we have a database that contains 405 170  170 texture images. To access our Corel database, suppose that a user submits a textual description ‘‘very fine ^ very directional ^ very regular,’’ as shown in Fig. 8a. Fig. 8b shows retrieval results for the query. The retrieved images are displayed in descending similarity order, from left to right and top to bottom; the similarity is shown below each retrieved image. If the user selects the second, fourth, and fifth

2263

images (in Fig. 8b) as relevant examples to refine the query, as shown in Fig. 8c, the retrieval results are improved, as shown in Fig. 8d. Suppose that the user submits a textual description ‘‘(directional _ blob-like) ^ very regular,’’ as shown in Fig. 9a. Fig. 9b shows retrieval results for the query. If the user selects the third, fifth, and eighth images (in Fig. 9a) as relevant examples to refine the query, as shown in Fig. 9c, the retrieval results are improved, as shown in Fig. 9d. The above two examples demonstrate the effective use of textual description queries and their logic compositions in our CBIR system. They also demonstrate the effective use of relevance feedbacks. Suppose that the user submits a visual example with the specified constrain ‘‘similar on coarseness and contrast,’’ as shown in Fig. 10a, to access our Corel database. Fig. 10b shows retrieval results for the query. If the user selects the fifth and eighth images (in Fig. 10b) as relevant examples to refine the query, as shown in Fig. 10c, the retrieval results are improved, as shown in Fig. 10d. The above example demonstrates the effective use of visual example queries on certain texture properties and relevance feedbacks in our CBIR system. Our CBIR system can help users to interpret texture images through high-level textual concepts. Given a texture image, the six Tamura features are first extracted. The membership value of each feature value in the linguistic term is then computed through the generated term set on the feature. The most feasible term to interpret the texture image is the one with the largest membership value. Fig. 11 shows four texture description examples. The effectiveness of our CBIR system is evaluated as the following three steps: Step 1: We submit each image in our VisTex database as a query to access our VisTex database and compute the retrieval efficiency by (Mehtre et al., 1997, 1998): n gT ¼

N n T

if N 6 T ; if N > T ;

where T is the number of retrieved images, n is the number of relevant images, and N

2264

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

Fig. 8. (a) A textual description; (b) retrieval results for the query; (c) relevant examples: the second, fourth, and fifth images in (b); (d) retrieval results after given examples in (c).

is the total number of relevant images in the database (N ¼ 9 in our experiments). Note that if N 6 T , gT becomes the recall measure of traditional information retrieval; if N > T , gT becomes the precision measure. Step 2: The relevant images are selected from the retrieved images to refine the query and access our VisTex database (i.e., a relevance feedback); the retrieval efficiency is again computed. The relevance feedback is iterated for three times. Step 3: We repeat Steps 1–2 by gradually increasing the number of the retrieved images, i.e., T , from 5, 10, 15, to 20. For each T ,

we compute the average retrieval efficiency for the query and each feedback as the effectiveness of our CBIR system. Table 2 summaries the effectiveness of our CBIR system. We find that the improvement between the query and the first feedback is the most noticeable; but subsequent feedbacks attain minor improvements. This indicates that, in our CBIR system, users can find the desired images with very few efforts. Moreover, the average retrieval time is less than 0.01 s for a textual description query; the average retrieval time is about 0.11 s for a visual example query. Processing a query in our CBIR system is very time-efficient.

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

2265

Fig. 9. (a) A textual description; (b) retrieval results for the query; (c) relevant examples: the third, fifth, and eighth images in (b); (d) retrieval results after given examples in (c).

5. Conclusions and further research A fuzzy logic CBIR system for finding textures is proposed in this study. A user can submit textual descriptions and/or visual examples to find the desired texture images. After the initial search, the user can give relevant and/or irrelevant examples to refine the query and improve the retrieval efficiency. According to our experimental results, contributions of this study are identified as follows. (1) The use of term set formulates a mapping from low-level Tamura features to high-level linguistic terms; it bridges the semantic gap between the two semantic levels. (2) The use of Tamura features and linguistic terms characterizes texture

properties at various semantic levels; and further, the use of linguistic terms achieves high-level texture manipulations, including retrieval and descriptions, through textual concepts. (3) The use of relevance feedbacks models the human perception subjectivity and improves the retrieval efficiency. (4) The proposed method of defining and computing the similarity between a query and a texture image are highly intuitive and computationally simple. Experimental results reveal our CBIR system is indeed effective. The retrieved images are perceptually satisfactory, and the retrieval time is very short. Here are several research directions we are pursuing. (1) Incorporate textual concepts of other

2266

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267

Fig. 10. (a) A visual example and a specified constrain; (b) retrieval results for the query; (c) relevant examples: the fifth and eighth images in (b); (d) retrieval results after given examples in (c).

Fig. 11. Texture description examples.

H.-C. Lin et al. / Pattern Recognition Letters 24 (2003) 2255–2267 Table 2 Effectiveness of our CBIR system gT

Query Feedback-1 Feedback-2 Feedback-3

n=T with T ¼5

n=N with T ¼ 10

n=N with T ¼ 15

n=N with T ¼ 20

61.77% 65.41% 65.25% 65.45%

51.03% 56.65% 58.19% 58.27%

59.97% 64.99% 66.47% 66.23%

65.46% 71.33% 72.70% 72.84%

image contents, such as colors, shapes, and the spatial layout, into our CBIR system. (2) Apply our CBIR system to those image databases that contain specific kinds of images, such as texture database for fabrics. (3) Design indexing structures or searching strategies to achieve more efficient image retrieval.

References Aigrain, P., Zhang, H.J., Petkovic, D., 1996. Content-based representation and retrieval of visual media: A state-of-theart review. Multimedia Tools Appl. 3 (3), 179–202. Ashley, J., Barber, R., Flickner, M.D., Hafner, J.L., Lee, D., Niblack, W., Petkovic, D., 1995. Automatic and semiautomatic methods for image annotation and retrieval in QBIC. Storage Retrieval Image Video Databases 2420, 24–35. Chang, S.F., Smith, J.R., Meng, H.J., Wang, H.L., Zhong, D., 1997. Finding images/video in large archives. D-Lib Mag. (Feb). Conners, R.W., Harlow, C.A., 1980. Towards a structural textural analyzer based on statistical methods. Comput. Graphics Image Process. 12 (3), 224–256. Del Bimbo, A., 1999. Visual Information Retrieval. Morgan Kaufmann Publishers. Flickner, M., Sawhney, H., Niblack, W., Ashley, J., Huang, Q., Dom, B., Gorkani, M., Hafner, J., Lee, D., Petkovic, D., Steele, D., Yanker, P., 1995. Query by image and video content: The QBIC system. IEEE Comput. Mag. 28 (9), 23–32. Goodrum, A.A., 2000. Image information retrieval: An overview of current research. Informing Sci. 3 (2), 63–67. Gupta, A., Jain, R., 1997. Visual information retrieval. Comm. ACM 40 (5), 70–79. Haralick, R.M., 1982. Image texture survey. In: Krishnaiah, P.R., Kanal L.N. (Eds.), The Hand-book of Statistics, vol. 2. pp. 399–415. Haralick, R.M., Shanmugam, K., Dinstein, I., 1973. Texture features for image classification. IEEE Trans. Systems Man Cybernet. 3 (6), 610–621.

2267

Idris, F., Panchanathan, S., 1997. Review of image and video indexing techniques. J. Visual Comm. Image Representation 8 (2), 146–166. Lin, H.C., Wang, L.L., Yang, S.N., 1997. Color image retrieval based on hidden Markov model. IEEE Trans. Image Process. 6 (2), 332–339. Lin, H.C., Wang, L.L., Yang, S.N., 1999. Regular-texture image retrieval based on texture-primitive extraction. Image Vision Comput. 17 (1), 51–63. Liu, F., Picard, R., 1996. Periodicity, directionality, and randomness: Wold features for image modeling and retrieval. IEEE Trans. Pattern Anal. Machine Intell. 18 (7), 722–733. Mehtre, B.M., Kankanhalli, M.S., Lee, W.F., 1997. Shape measures for content-based image retrieval: A comparison. Information Process. Management 33 (3), 319–337. Mehtre, B.M., Kankanhalli, M.S., Lee, W.F., 1998. Contentbased image retrieval using a composite color-shape approach. Information Process. Management 34 (1), 109–120. Nadler, M., Smith, E.P., 1993. Pattern Recognition Engineering. John Wiley & Sons. Niblack, W., Barber, R., Equitz, W., Flickner, M., Glassman, E., Petkovic, D., Yanker, P., 1993. The QBIC Project: Querying images by content using color, texture, and shape. Storage Retrieval Image Video Databases 1908, 173–181. Niblack, W., Flickner, M., 1993. Find me the pictures that look like this: IBMÕs QBIC Image Query Project. Adv. Imaging 8 (2), 32–35. Pentland, A., Picard, R.W., Sclaroff, S., 1994. Photobook: Tools for content-based manipulation of image databases. SPIE Storage Retrieval Image Video Databases 2185, 34–47. Rui, Y., Huang, T.S., Chang, S.F., 1999. Image retrieval: Current techniques, promising directions, and open issues. J. Visual Comm. Image Representation 10 (1), 39–62. Rui, Y., Huang, T.S., Ortega, M., Mehrotra, S., 1998. Relevance feedback: A power tool for interactive contentbased image retrieval. IEEE Trans. Circuits Syst. Video Technol. 8 (5), 644–655. Shapiro, L.G., Stockman, G.C., 2001. Computer Vision. Prentice Hall. Smeulders, A.W.M., Worring, M., Santini, S., Gupta, A., Jain, R., 2000. Content-based image retrieval at the end of the early years. IEEE Trans. Pattern Anal. Machine Intell. 22 (12), 1349–1380. Tamura, H., Mori, S., Yamawaki, T., 1978. Texture features corresponding to visual perception. IEEE Trans. Systems Man Cybernet. 8 (6), 460–473. Tuceryan, M., Jain, A.K., 1998. Texture analysis. In: Chen, C.H., Pau, L.F., Wang, P.S.P. (Eds.), The Handbook of Pattern Recognition and Computer Vision, 2/e. World Scientific, pp. 207–248. Zimmermann, H.J., 1991. Fuzzy Set Theory and Its Applications, 2/e. Kluwer Academic Publishers.

Finding textures by textual descriptions, visual ... - Semantic Scholar

that LinХs system deals with regular textures only. Tamura et al. (1978) proposed a texture representation based on psychological studies of human perceptions.

674KB Sizes 5 Downloads 278 Views

Recommend Documents

combining textual and visual clusters for semantic ...
retrieve images using semantic keywords and auto-annotate images based on ... can learn (about) natural language from training data such as document ...

Multi-Sentence Compression: Finding Shortest ... - Semantic Scholar
Proceedings of the 23rd International Conference on Computational ... sentence which we call multi-sentence ... tax is not the only way to gauge word or phrase .... Monday. Figure 1: Word graph generated from sentences (1-4) and a possible ...

Automated Textual Descriptions for a Wide Range of ...
Abstract. Presented is a hybrid method to generate textual descrip- tions of video based on actions. The method includes an action classifier and a description generator. The aim for the action classifier is to detect and classify the actions in the

Semantic Queries by Example - Semantic Scholar
Mar 18, 2013 - a novel method to support semantic queries in relational databases with ease. Instead of casting ontology into rela- tional form and creating new language constructs to express ...... uni-karlsruhe.de/index_ob.html. [19] OTK ...

Semantic Queries by Example - Semantic Scholar
Mar 18, 2013 - Finally, we apply the query semantics on the data to ..... mantic queries involving the ontology data are usually hard ...... file from and to disk.

Effectively finding relevant web pages from linkage ... - Semantic Scholar
semistructured data models and retrieval algorithms, object-oriented programming methods, and techniques. Yanchun Zhang received the PhD degree in.

Netting the evidence: finding pearls, not sewage - Semantic Scholar
information management principles of focusing on the original ... “The internet is transforming health care. It is ... to its packaging, its mode of administration, its.

Finding k-Dominant Skylines in High Dimensional ... - Semantic Scholar
large number of users, the data set being handled is obvi- ously a high-dimensional .... We provide some theo- retical analysis on the reason for its superiority.

Netting the evidence: finding pearls, not sewage - Semantic Scholar
Fax: (44) 114 272 4095. Email: ... e-mail; the channel by which PubMed, the free version of ..... above and you start to get the degree of precision required to ...

A Randomized Algorithm for Finding a Path ... - Semantic Scholar
Dec 24, 1998 - Integrated communication networks (e.g., ATM) o er end-to-end ... suming speci c service disciplines, they cannot be used to nd a path subject ...

New Methods in Finding Binary Constant Weight ... - Semantic Scholar
Master's Thesis. Date/Term: .... When a code is used to transmit information, the distance is the measure of how good the ... the best we can do is to find generalized bounding formulas. .... systems, overlap the theory of constant weight codes.

Netting the evidence: finding pearls, not sewage - Semantic Scholar
on using the Internet to answer clinical questions ..... the same materials whereas in the past, a single route .... Health Info Libr J 2005; 22 suppl 1:66-9. 20.

Visual Steering for Program Debugging Abstract 1 ... - Semantic Scholar
As software systems become more complex and must handle ever ... application to program visualization where it could prove quite beneficial in reducing debugging time. The ... These interaction philosophies can be very important to the development of

Motion integration and postdiction in visual ... - Semantic Scholar
176, 343 (1977); R. Wetts, G. N. ... A series of psychophysical experiments yields data inconsistent ... 17 MARCH 2000 VOL 287 SCIENCE www.sciencemag.org.

Graphical Models of the Visual Cortex - Semantic Scholar
chain. The activity in the ith region is influenced by bottom-up feed-forward data xi−1 and top-down .... nisms which combine a bottom-up, small-image-patch, data-driven component with a top-down ..... Optimal Statistical Decisions. New York: ...

Efficiently Training A Better Visual Detector With ... - Semantic Scholar
[4] proposed Float-. Boost for a better detection accuracy by introducing a back- ward feature elimination step into the AdaBoost training procedure. Wu et al. [5] used forward feature selection for fast training by ignoring the re-weighting scheme i

Visual perception of the relative phasing of human ... - Semantic Scholar
phase variability were used to generate computer displays of spheres oscillating either side to side in .... IIci and shown on an Apple 13-in. color display.

Subject-Adaptive Steady-State Visual Evoked ... - Semantic Scholar
command Brain-Computer Interface (BCI) based on steady- state visual evoked ... research grant GOA 10/019, AC and AR are supported by IWT doctoral grants, MvV is ... experiments we have used a laptop with a bright 15,4”. LCD screen ...

Modeling Drivers' Visual Attention Allocation While ... - Semantic Scholar
tracker system (Version 3.0.1), which consisted of three Sony XC HR50 ... of the driver for the eye tracker. ...... Transportation Research Part F, 5, 87–97. Vidulich ...

Syllabic length effects in visual word recognition ... - Semantic Scholar
Previous studies on the syllable-length effect in visual word recognition and naming ..... as subjects for course credit, 16 in the Experiment 1 (naming) and 40 in ... presented in isolation on the center of the display screen of a Pentium computer.

Subject-Adaptive Steady-State Visual Evoked ... - Semantic Scholar
ers [9]. This means that, when the subject is looking at stimuli flickering at the frequency f1, the ... experiments we have used a laptop with a bright 15,4”.

Influence of Depth Cues on Visual Saliency - Semantic Scholar
sion research, such as autonomous mobile systems, 3D content surveillance ...... Seo, H., Milanfar, P.: Static and space-time visual saliency detection by self-.

Efficiently Training A Better Visual Detector With ... - Semantic Scholar
Experiments in the domain of highly skewed data distri- butions, e.g. ...... NICTA is funded by the Australian Government as represented by the. Department of ... the Australian Research Council through the ICT Centre of Excellence program.