Tagging Tags ∗

Kuiyuan Yang

Xian-Sheng Hua

Univ. of Sci. & Tech. of China Hefei, Anhui, 230027 China

Microsoft Research Asia Beijing, 100190 China

[email protected]

[email protected] Hong-Jiang Zhang

Meng Wang

AKiiRA Media Systems Inc Palo Alto 94301, CA, USA

Microsoft Adv. Tech. Center Beijing, 100190 China

[email protected] [email protected] ABSTRACT

Keywords

Social image sharing websites like Flickr have successfully motivated users around the world to annotate images with tags, which greatly facilitate search and organization of social image content. However, these manually-input tags are far from a comprehensive description of the image content, which limits effectiveness of the tags in content-based image search. In this paper, we propose an automatic scheme called tagging tags to supplement semantic image descriptions by associating a group of property tags with each existing tag. For example, an initial tag “tiger” will be further tagged with “white”, “stripes” and “bottom-right” along three tag properties: color, texture and location, respectively. In the proposed scheme, a lazy learning approach is first applied to estimate the corresponding image regions of each initial tag, and then a set of property tags, which involve six exemplary property aspects including location, color, texture, shape, size and dominance, are derived for each tag according to the content of the regions and the entire image. These tag properties enable much more precise image search especially when certain tag properties are included in the query. The results of the empirical evaluation show that tag properties remarkably boost the performance of social image retrieval.

Tagging Tags, image retrieval

1. INTRODUCTION In recent years, with the rapid development of digital cameras and photo sharing websites, hundreds of millions of photos are uploaded, tagged by more than 8.5 million registered users, and these numbers are keep growing [11]. These tags make the massive photos semantically searchable [12, 13]. However, there is still a gap between the manually-input tags and the users’ queries. One of the primary reasons is that the set of so-called basic-level tags [7] are much more frequently used when tagging the images. The basic-level theory [10] states that terms can be cognitively structured in a hierarchical system with three different levels of specificity: (1) the superordinate level, (2) the basic level and (3) the subordinate level. The basic level often contains terms that are mostly demonstrative, but not specific. Rorissa investigates user behaviors regarding image indexing, he finds out that the basic level is generally used for the description of a single object and the superordinate level for description of an object group: “When describing images, people mainly made references to objects (mainly named at the basic level) in images much more than other types of attributes” [9]. Goodrum and Spink also analyze the searching behavior during the retrieval of visual information and observe that often, users initially search for basic-level terms, but then restrict the search results using so-called “refiners” [3]. Thus the more specific, subordinate level queries are employed during image search, for example, terms like “red” and “round” may serve to refine a general term such as “balloon” into a more specific search request. To reduce the gap between the tags and queries, in this paper we propose a so-called tagging tags scheme, which automatically adds a set of property tags to each of the existing initial tag. Six exemplary tag properties will be studied in this paper, including location, color, shape, texture, size and dominance. Tags provided by users can be classified into two categories, one is context-related tags, and the other is content-related tags. Context-related tags are not directly related to the content of the image, while content-related tags generally only tell what appear in the image, but not much information about where and how they appear, for example, in Fig. 1, all the four images are tagged with “balloon”, but they are with different locations, colors, shapes

Categories and Subject Descriptors H.3.1 [Information Storage and Retrieval]: Content Analysis and Index

General Terms Algorithms, Experimentation ∗This work was performed when Kuiyuan Yang was visiting Microsoft Research Asia as a research intern.

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. MM’10, October 25–29, 2010, Firenze, Italy. Copyright 2010 ACM 978-1-60558-933-6/10/10 ...$10.00.

619

(a)

(b)

(c)

instance’s tag and intuitively can be converted into a multiinstance learning problem. Diverse Density is a general framework proposed by Maron for solving multi-instance problem [5]. Diverse Density (DD) at a point in the instance feature space is defined as a measure of how many different positive bags have instances near that point, and how far the negative instances are from that point. An optimization algorithm such as gradient descent approach is used to search for the point achieves maximal DD, and once a point with maximum DD is found, a new instance can be classified according to the distance between the instance and the maximum DD point. In DD, one and only one point is taken as a general model for each class. However this is actually not suitable for tag to region task, since one tag may correspond to multiple local maximum points in the feature space, for example one tag frequently corresponds to more than one region in an image, e.g., “car” in an image is segmented into wheels, windows and body, each part corresponds to one maximum DD in the feature space. Instead of learning a general model for each tag, we adopt a lazy learning approach based on Diverse Density, that is Lazy Diverse Density. Since each point’s DD measures its positive degree, we directly estimate the DD of each instance, and each instance’s tag is determined according to its DD. Due to each instance has its own best descriptive features, e.g., “flower” is well described by color while “zebra” is well described by texture etc, we map each instance into three different feature spaces (color, shape and texture), and compute its DD in each feature space, then the feature space with the highest DD is regarded as best description for the instance. Finally, the instance’s tag is determined by comparing DDs computed with respect to tags from its bag. The Lazy Diverse Density approach for tag to region is detailed as follows. Let D = {Bi , Ti }N i=1 denotes the tagged image set, where N is the total image number. Bi = {Bi1 , Bi2 , . . . , Bini } represents an image, where Bij is an over-segmented image patch by using graph-based segmentation algorithm in [2], and Ti = {Ti1 , Ti2 , . . . , Timi } is the tag set associated with the image Bi . For each image patch, we extract three types of features, denoted by F ={color, shape, texture}. For an instance x ∈ B, T is the tag set associates with B, its DD with respect to tag t ∈ T in feature space f is defined as following:

(d)

Figure 1: An example to illustrate the descriptive information is missed by tags. All the four images are tagged with “balloon”, but they are with different locations, colors, shapes and sizes. The pink balloon in (d) is even hardly visible in the image. All this information is hidden behind the tag “balloon”. and sizes, the pink balloon in (d) is even hardly visible in the image. All this information is hidden behind the tag “balloon”. Property tags are used to restrict tags from basic level into subordinate level. In this paper, we only define property tags for content-related tag based on the information from image content, while inferring properties of context-related tags will be our future work and out of this paper’s scope. Images indexed by these property tags can support query with “refiners” like “find a round green balloon on the topright corner”, and Fig. 1(c) will be returned as a matched result. However, automatically assigning tag properties is a nontrivial task. As most of the tag properties are certain characteristics of a target object, the first step of detection a tag property is to find the regions in the image which correspond to the target tag. In our scheme, we propose an efficient approach called Lazy Diverse Density to estimate the corresponding image region of each tag, and then the property tags are derived according to the content of the image region. To the best of our knowledge, this work represents the first attempt towards automatically expand tags with descriptive information. The main contributions of this paper can be summarized as follows: (1) Proposed a new tagging tags scheme to mine more descriptive information for the initial tags. (2) Proposed a Lazy Diverse Density approach for fast tag-region matching. (3) More precise image retrieval is supported based on the property tags. The rest of the paper is organized as follows. We first introduce our tagging tags scheme in Section 2 and introduce image search with tag properties in Section 3. Then we give experimental results in Section 4. Finally, we conclude the paper in Section 5.

2.

X DD(x, t, f ; D) ≈

1≤i≤N,t∈Ti

´ X ³ 1 − δ[xf ∼ Bif ]

1≤i≤N,t6∈Ti

N

(1)

where δ[expression] is set to 1 when the expression is true, and 0 otherwise, xf ∼ Bif means bag Bi and x is “near” in feature space f , where “near” means x has any instance of Bi as its k-Nearest Neighbor or any instance of Bi cites x as its k-Nearest Neighbor in feature space f . The feature space with maximum DD is regarded as best describes the instance, then DD of instance x with respect to tag t is defined as,

TAGGING TAGS

The tagging tags scheme mainly consists of two steps: tag to region and tagging property tags. Tag to region is to estimate each tag’s corresponding image region through Lazy Diverse Density. Tagging property tags is based on the estimated image regions in the first step. In this section, we will detail Lazy Diverse Density approach for tag to region and tagging property tags, respectively.

2.1

δ[xf ∼ Bif ] +

Overview of Problem and Solution

DD(x, t; D) = max DD(x, t, f ; D) f ∈F

By viewing each image as a bag, each of which contains a number of instances corresponding to regions obtained from image segmentation, then tag to region is to estimate each

(2)

The tag of instance x is set as the tag with the highest DD. Due to tags are incomplete, some instances have no cor-

620

2.2.4 Tagging Color, Texture and Shape top-left

above

top-right

left

center

right

bottom-left

below

bottom-right

Color, texture and shape are the three most widely used visual descriptors. We define 11 basic color tags, 13 common texture tags and 5 common shape tags, denoted by Tc ={black, blue, brown, green, orange, pink, purple, red, white, yellow}, Tt ={stripes, spots, skin, furry, clouds, water, grass, tree, bricks, check, tough skin, rock, wood} and Ts ={round, rectangle, triangle, diamond, heart} respectively. These property tags for each target tag is derived according to the estimated image region. In the following, we will detail the tagging process of color tags, the tagging process of texture tags and shape tags are omitted since they analogous to the color tags. Intuitively, color tag can be set by comparing the RGB value of the image region with each color tag’s definition in RGB color space. However these definitions are under ideal lighting on a neutral background may greatly differ from the color in the real world. To make the color tag more consistent with real world, we propose to learn the color tags from real-world images. Similar to tag-to-region task, tagging color is to choose one color tag from the tag set Tc for each instance, so we adopt the same as Lazy Diverse Density for the task. We select 50 true positive images for each of the 11 color names from search results returned by an image search enc gine, and denote the collected image set as C = {Bi , ci }N i=1 , where Bi represents a bag obtained the same as bags used in tag to region and ci ∈ Tc is its color tag. For target tag to , denote its corresponding image region as x. DD(x, c, color; C) is the Diverse Density of x computed with respect to color tag c ∈ Tc in the color feature space using the bags of image set C, it is computed using Eq. 1. Then color tag for target tag to is set similar to Eq. 3,

Figure 2: Image is partitioned into 3-by-3 equal grids, each grid corresponds to one location tag. responding tag in the bag’s tag set, so instance with highest DD below a threshold thresh is assigned with “background”, ( t(x) =

background, if max DD(x, t; D) ≤ thresh t∈T

arg max DD(x, t; D), else

(3)

t∈T

After each image region is assigned with tag, adjacent regions with the same tag are merged together.

2.2

Tagging Property Tags

After the corresponding image region is estimated for each tag, then other property tags can be derived according to the content of the corresponding image region. For each target tag, it may correspond to several regions1 in one image, we use the one with highest DD to derive its property tags.

2.2.1 Tagging Location Location tag tells where the tag appears in the image. Different location descriptors can be used here, such as coordinate and orientation. In current implementation, each image is partitioned into 3-by-3 equal grids, and each grid corresponds to one location tag as shown in Fig. 2 and 9 location tags are defined in total. The location tag for each initial tag is given according to which grid the centroid of its corresponding image region dropped into.

2.2.2

( c(to ) =

N A, if max DD(x, c, color; C) ≤ threshc c∈Tc

(6)

arg max DD(x, c, color; C), else c∈Tc

where N A means not set color tag for this target tag if the DD below threshold threshc .

Tagging Size

Size measures how big the tag-specified regions appear in the image, which is meaured by the sum of area-ratios of all regions tagged with tag t, X S(t) = Ap (4)

3. IMAGE SEARCH WITH PROPERTY TAGS To prove the effectiveness of tag properties, in the section, we introduce a simple BM25-based [8] image search scheme that takes advantage of property tags. Each image is represented by its original tags and also the property tags, which can be regarded as a short document. Okapi BM25 [8] is a bag-of-words retrieval function that ranks a set of documents based on the query terms appearing in each document. The BM25 score of a image Bj with respect to a keyword q is defined as,

p∈Pt

where Pt denotes all the regions tagged with t, and Ap denotes the area-ratio of region p.

2.2.3 Tagging Dominance Diverse Density of an image region with respect to the tag can be regarded as its relevance degree to the tag. Dominance measures how significant a tag appears in one image, and it is related to both the relevance and size of the tag-specified regions, which can be measured by the sum of Diverse Densities weighted by the corresponding area-ratios, X D(t) = DD(xp , t; D)Ap (5)

BM 25(Bj , q) = IDF (q)

tf (Bj , q)(k1 + 1) tf (Bj , q) + k1 (1 − b +

b|Bj | avgdl

(7)

Where tf (Bj , q) is q’s term frequency in the image Bj , |Bj | is the total number of tags associate with Bj , and avgdl is the average tag number of the image collection. k1 and b are free parameters, set as k1 = 2 and b = 0.75 in this paper. IDF (q) is the IDF (inverse document frequency) weight of query term q. Given a query Q, containing keywords q1 , . . . , qn , the BM25

p∈Pt

where Pt denotes all regions tagged with t. 1 Adjacent regions corresponding to the same tag are merged into one before this step.

621

score of an image Bj is: n X

BM 25(Bj , qi )

(8)

Average Prec@X

score(Bj , Q) = BM 25(Bj , Q) +

i=1

where we add the first term in the r.h.s to take some compound queries into consideration, e.g., “red apple”. After obtaining the BM25 scores, images are sorted with a nonincreasing order. Since users often view few result pages when perform search [4], images relevant to user queries should be ranked as high as possible. We evaluation the search results by employing P rec@X, #relevant images in top X P rec@X = X

4.

0.3 0.2

0 5

10

15

20

25

30

35

40

X

Figure 3: Average Prec@X comparison with varied X of image search using different methods.

(9)

5. CONCLUSIONS AND FUTURE WORK

EXPERIMENTS Data Set Description

We conduct tagging tags on NUS-WIDE dataset [1], which contains 269,648 images that are collected from Flickr, there are 425,059 tags associated with these images originally. Due the noisy and synonyms in the original tags, we keep tags that belong to the “physical entity” group in the WordNet, and change plural to singular. Synonyms are grouped using WordNet synsets and tags appear with too low frequencies are filtered out. Finally 601 tags are kept after these processes. Images with no tag are not used and thus we obtain a subset with 66,015 images, and in average 3.48 tags associated with each image. Then images are segmented into image patches, patches that are smaller than 1% of the original image are ignored. This results in 1,108,285 legible patches in total. For each image patch, we extract 125-dimensional color histogram features, 80-dimensional edge histogram features (edge histogram is one kind shape feature) and 256-dimensional LBP texture features[6]. The neighbor number k used for computing Eq. 4 is empirically set to 50 for all instances, thresholds thresh, threshc , thresht and threshs are empirically set to 0.8.

4.2

0.4

0.1

In this section, we systematically evaluate the effectiveness of property tags on a publicly available dataset.

4.1

Image Search with Tag Properites Image Search with Original Tags

0.5

Image Search

We conduct image search on the 66,015 images described in Section 4 to verify the effectiveness of property tags. We first select 58 basic queries, including scenes and objects. For each basic query, we add meaningful modifiers to form specific queries (e.g., “girl on the left”, “red balloon”, etc.) and 804 queries are formed in total. We compare the image search results based on the following indexing methods: (1) Image Search with Original tags (ISOriginal), i.e., index the images with original tags. (2) Image Search with tag properties (ISProperty), i.e., index the images with both the original tags and the property tags added by our method. To quantitatively compare the image search results, we invited 10 subjects to manually label the relevance of the top 40 images of each query and each method. The average P rec@X results with varied X are illustrated in Fig. 3, which demonstrate that our method achieves much better search performance.

622

This paper proposed a tagging tags scheme aiming at expanding existing tags with a set of property tags to supplement the missed descriptive information. With the popularity of the photo sharing websites, the community-contributed images with tags are much easier to obtain, the keyword query based semantic image search can greatly benefit from applying our proposed technique for adding property tags. Experiments also has proven that search with property tags significantly improves the search performance specially for queries with specific modifiers.

6. REFERENCES

[1] T.-S. Chua, J. Tang, R. Hong, H. Li, Z. Luo, and Y.-T. Zheng. NUS-WIDE: A Real-World Web Image Database from National University of Singapore. In CIVR, 2009. [2] P. Felzenszwalb and D. Huttenlocher. Efficient graph-based image segmentation. IJCV, 2004. [3] A. Goodrum and A. Spink. Image searching on the excite web search engine. Information Processing & Management, 2001. [4] B. Jansen and A. Spink. How are we searching the World Wide Web? A comparison of nine search engine transaction logs. Information Processing & Management, 2006. [5] O. Maron and T. Lozano-P´ erez. A framework for multiple-instance learning. In NIPS, 1998. [6] T. Ojala, M. Pietik¨ ainen, and T. M¨ aenp¨ a¨ a. Gray scale and rotation invariant texture classification with local binary patterns. In ECCV, 2000. [7] I. Peters. Folksonomies: Indexing and Retrieval in the Web 2.0. De Gruyter, 2009. [8] S. Robertson, S. Walker, S. Jones, M. Hancock-Beaulieu, and M. Gatford. Okapi at TREC-4. In the Fourth Text Retrieval Conference, 1996. [9] A. Rorissa. User generated descriptions of individual images versus labels of groups of images: A comparison using basic level theory. Information Processing & Management, 2008. [10] E. Rosch, C. Mervis, W. Gray, D. Johnson, and P. Boyes-Braem. Basic objects in natural categories. Cognitive psychology, 1976. [11] B. Sigurbj¨ ornsson and R. van Zwol. Flickr tag recommendation based on collective knowledge. In WWW, 2008. [12] K. Yang, M. Wang, X.-S. Hua, and H.-J. Zhang. Social image search with diverse relevance ranking. In Advances in Multimedia Modeling, 2010. [13] K. Yang, M. Wang, and H.-J. Zhang. Active tagging for image indexing. In ICME, 2009.

Tagging tags

AKiiRA Media Systems Inc. Palo Alto ..... Different location descriptors can be used here, such as co- .... pound queries into consideration, e.g., “red apple”. After.

1MB Sizes 0 Downloads 274 Views

Recommend Documents

Advent tags printable.pdf
Page 1 of 8. Make a gingerbread house. Put up the. Christmas tree. Sing Christmas. carols together. Visit an old age home. Play games or sing them. some of the carols you've. learned. Purge toys and donate. them to a thrift store. (A. great way to ma

Mobile App Tagging
Mobile app markets; app tagging; online kernel learning. 1. INTRODUCTION ... c 2016 ACM. .... and regression [8], multimedia search [23], social media, cy-.

man-106\all-tags-xxnx.pdf
The book All Tags Xxnx supplies the very best encounter as well. as lesson to take, ... PDF Ebook : Xxnx Mobile Phone. 4. ... PDF Ebook : Dell Service Tags. 12.

Animal ear tags and applicators therefor
Apr 29, 1971 - flange 17 ?tting into the groove 15 for the spike to become an integral part of the component 10. The second component 18 is formed with the ...

HiPSTAS Poetry Tagging -
2005 article on “Feminization of Voice and. Communication.” http://speech-language-pathology- audiology.advanceweb.com/Article/Feminization-of-Voice-and-.

Dog Tags 1.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Dog Tags 1.pdf.

CreativeIndex-Tags-Red-Green.pdf
There was a problem loading more pages. Retrying... CreativeIndex-Tags-Red-Green.pdf. CreativeIndex-Tags-Red-Green.pdf. Open. Extract. Open with. Sign In.

Dog Tags 2.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Dog Tags 2.pdf. Dog Tags 2.pdf. Open. Extract. Open with. Sign In.

TagBase - Tag database for fish tags - GitHub
TagBase is a relational database application for the management of output files from manufacturer's post-recovery or post-reporting software. TagBase ...

frozen olaf tags printable.pdf
Page 1 of 1. frozen olaf tags printable.pdf. frozen olaf tags printable.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying frozen olaf tags printable.pdf.

Animal ear tags and applicators therefor
Apr 29, 1971 - “Dudtag Ear Tags", Brochure of Delta Plastics, Ltd. published about ..... and said spike being adapted to receive a support rod for positioning ...

Gift tags by IMS.pdf
Loading… Whoops! There was a problem loading more pages. Whoops! There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Gift tags by IMS.pdf. Gift tags by IMS.pd

active tagging for image indexing
quantized since there should be enormous (if not infinite) potential tags that are relevant to ... The concurrence similarity between ti and tj is then defined as. W. T.

CreativeIndex-Tags-Gold-Black.pdf
CreativeIndex-Tags-Gold-Black.pdf. CreativeIndex-Tags-Gold-Black.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying ...

SAW Spread Spectrum RFID Tags and Sensors
the SAW CDMA tag is wireless and passive, while the Si tag is an active tag that requires ... still low cost and has similar advantages to the CDMA approach, will ...

Dog Tags 1.pdf
Sign in. Page. 1. /. 1. Loading… Page 1 of 1. Page 1 of 1. Dog Tags 1.pdf. Dog Tags 1.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying Dog Tags 1.pdf. Page 1 of 1.

People Tagging & Ontology Maturing: Towards ...
Combine Web 2.0-style bottom-up processes with organizational top-down processes. • Competence management without an agreed vocabulary (or ontology) ...

active tagging for image indexing
many social media websites such as Flickr [1] and Youtube [2] have adopted this approach. .... For Flickr dataset, we select ten most popular tags, including.

Children's Market Valet Tagging Consignor Agreement - Spring ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Children's Ma ... ring 2018.pdf. Children's Ma ... ring 2018.pdf. Open.

Context-Dependent Fine-Grained Entity Type Tagging
Dec 3, 2014 - Dan Gillick, Nevena Lazic, Kuzman Ganchev, Jesse Kirchner, David Huynh. Google Inc. Abstract. Entity type tagging is the task of assign- ing category labels to each ...... Machine Learning Research, 7(1). [Doddington et al.2004] George

Gas Leaks Tagging Project -
We want to set up a core group of advisors and “worker bees” to help plan and recruit tag team leaders and tagging teams. ... accountable for their infrastructure. It is estimated there are 100+ ... An ideal outcome of this initiative: pressure t

DIY Printable Groundhog Day Tags Toppers Psmerker Designs ...
DIY Printable Groundhog Day Tags Toppers Psmerker Designs 2016.pdf. DIY Printable Groundhog Day Tags Toppers Psmerker Designs 2016.pdf. Open.