Unsupervised Image Categorization and Object Localization using Topic Models and Correspondences between Images David Liu and Tsuhan Chen Department of Electrical and Computer Engineering Carnegie Mellon University, Pittsburgh, PA, USA

Abstract Topic models from the text understanding literature have shown promising results in unsupervised image categorization and object localization. Categories are treated as topics, and words are formed by vector quantizing local descriptors of image patches. Limitations of topic models include their weakness in localizing objects, and the requirement of a fairly large proportion of words coming from the object. We present a new approach that employs correspondences between images to provide information about object configuration, which in turn enhances the reliability of object localization and categorization. This approach is efficient, as it requires only a small number of correspondences. We show improved categorization and localization performance on real and synthetic data. Moreover, we can push the limits of topic models when the proportion of words coming from the object is very low.

1. Introduction This paper addresses the problem of unsupervised object localization and image categorization. The problem is challenging because objects of the same type may appear in different locations, scales, or under occlusions and deformations. In [19][17][6][8], topic models were used to model the appearance of each object class by learning the typical proportions of visual words and have shown promising results. In this paper we will examine a number of extensions on topic models made previously in literature for the task of unsupervised image categorization and object detection in the presence of background clutter. Many of these extensions have not been analyzed in a comparative manner. Specifically, we will be investigating an approach that does not use location information at all, and some other approaches that model the location of patches. We propose a combination of a topic model approach and correspondence-based approach, and show significant improvements over current topic models that model the lo-

Figure 1. Correspondences between patches across images (red lines on the left) provide strong information of the object configuration. This information, shown at the bottom right as a “reward map”, is incorporated into a topic model in our approach to enhance object localization and image categorization.

cation of patches. Our approach is advantageous over existing topic model approaches due to its non-parametric representation of the object’s spatial configuration (see Figure 1). With this non-parametric representation, we can obtain better estimates of the labels for each patch, thereby achieving significant better localization and categorization results. We will give an overview of topic models in Section 2. We will then detail our proposed method in Section 3, and conduct a series of experiments with synthetic and real data in Section 4; in particular, we will analyze the performance extensively with different distortions and data sets, and results on search engine image re-ranking.

2. Topic models Topic models such as Probabilistic Latent Semantic Analysis (PLSA)[11], were originally used in the text understanding community for unsupervised topic discovery. In computer vision, topic models have been used to discover object classes, or topics, from a collection of unlabeled images. As a result, images can be categorized according to the topics they contain. In the context of unsupervised object detection, the object of interest and the background

clutter are the two topics. Visual words (or textons) [16] are vector quantized local appearance descriptors from patches or edge curves. Objects can be represented as collections of visual words. Images consist of one or more object classes, which often have significant background clutter. Following the notations used in the text understanding community, w ∈ W = {w1 , ..., w|W | } is the visual word associated with a patch, z ∈ Z = {z1 , ..., z|Z| } is a hidden variable that represents the topic associated with a patch, and d ∈ D = {d1 , ..., d|D| } is the index of the image associated with a patch. In the following, we will investigate several existing topic models.

2.1. A1: No Location or Shape Info PLSA has been used as an appearance-based model [19][17]; no location or shape info is used. PLSA assumes the joint distribution of d, w,and z can be written as P (d, w, z) = P (d)P (z|d)P (w|z). PLSA is known for its capability of handling polysemy: if a visual word w is observed in two images di and dj , then the topic associated with that word can differ in di and dj : arg max P (z|di , w) can be different from arg max P (z|dj , w). In other words, PLSA allows a visual word to have different meanings in different images. When w is merely an appearance descriptor and contains no spatial information, the model does not care about the spatial ordering of the w’s. This is problematic when spatial information is an important cue for recognition. Below we will discuss modified versions of PLSA that consider the spatial ordering of the w’s in different ways.

2.2. A2: Spatial Location In [14], the topic model specifies where the object is more likely located. Let r denote the location of a patch. For a patch in image d with appearance w and location r, the joint distribution P (d, w, r) has the form P (d, w, r, z) = P (d)P (z|d)P (w|z)P (r|z). P (r|z) is a spatial distribution that models where a patch with topic z is more likely to occur. This model can be useful in modeling pedestrians for instance less likely to walk in the sky. Including location information more or less solves the ambiguity mentioned in A1, that is, the spatial ordering of w’s now has an impact on the discovery of topics, even when the topic appearance has large ambiguity (large overlap between the distributions P (w|zi ) and P (w|zj ), i = j). The spatial distribution P (r|z) uses the same parameters across all documents, hence it is a global location model, and it is not translation nor scale invariant.

r d

z w

Figure 2. The graphical model. The outer plate indicates the graph is replicated for each image, and the inner plate indicates the replication of the graph for each patch. The topic variable z is hidden.

2.3. A3: Spatial Clustering Recent work by [8][15] are based on the assumption that an object normally consists of patches that co-exist tightly rather than scattered around loosely. In [15], the location and scale of the object is hypothesized through a spatial distribution P (r|d, z), and the joint distribution is P (d, w, r, z) = P (d)P (z|d)P (w|z)P (r|d, z). The algorithm for finding the topic and object location and scale can be viewed as performing joint spatial and appearance clustering. Different than A2, the parameters for the spatial distribution P (r|d, z) are estimated per image, hence it provides translation and scale invariance. Notice that this approach only models the spatial clustering behavior of patches but it does not model the relative position or the ordering between patches.

2.4. Common framework Before we describe our model, let us review the commonalities among A1, A2 and A3. They all consist of image characteristic, P (z|d), and topic appearance, P (w|z), as shown in the graphical model in Figure 2 by d → z and z → w. In addition, A2 and A3 assign an additional feature r for each patch that specifies its location. A patch is then “rewarded” according to how much its location r agrees with the spatial model. The amount of reward is specified by the distribution P (r|z) in A2, or P (r|d, z) in A3. The graphical model in Figure 2 illustrates the more general case of P (r|d, z); if r is assumed to be independent of d when given z, then P (r|d, z) reduces to P (r|z).

3. Reward by Correspondence While A2 and A3 make use of spatial information, neither of them considers the spatial configuration of patches coming from the object of interest, which are normally far more consistent than patches coming from background clutter. In the context of unsupervised object detection, it is the consistency of patches across images that tells an object apart from the background clutter: similar objects that appear repeatedly in the data set will demonstrate a consistent spatial configuration, while the patches from random background clutter lack consistent spatial configuration. Sim-

Figure 3. An image and three exemplar images. Red lines indicate correspondences found between patches.

ilarly, in the context of image categorization, same object classes share similar spatial configurations, which are distinct for every object class. Our intent is to extend the topic model framework to incorporate information about spatial configuration. Rather than building a shape model [20][7], we will exploit the fact that similar objects in different images are more likely to have strong correspondences and extend the topic model to include this extra piece of information. Correspondence-based object recognition has been in the literature for nearly forty years. Even though computing correspondence is computationally expensive, it is still popular, because of its promising performance. Recent work by [3][4] [9] [13] use correspondence as a central element in their object recognition framework. However, their model and learning algorithm differ substantially from what is proposed here. Our use of correspondence is to provide a nonparametric representation of the location of the consistent patches. By using correspondence methods that take into account the spatial distortion of a correspondence and allow partial matching, the more matches a patch has, the better chances the patch belongs to a foreground object, as opposed to background clutter. This piece of information is employed by the topic model in the form of an extra feature, or reward. More precisely, the reward for a patch is high when this patch is repeatedly matched against other patches in other images. On the other hand, patches from random clutter are less likely to find as many matches, which results in lower rewards. This is precisely the notion that objects of interest normally show higher consistency across images and it is the consistency that tells an object apart from background clutter or other objects. The more good matches a patch gets, the higher its reward, and the resulting reward map, as illustrated in Figure 1, is a non-parametric representation of the location of the consistent patches. It is important to use correspondence methods that respect both appearance and geometric costs. Correspondence methods such as [4] [13] are among these. They measure the cost of a correspondence by observing how similar feature points are to their corresponding feature points, and how much the spatial arrangement of the feature points is changed. These methods also allow outliers to be excluded

from the correspondence. Computing correspondences between all pairs of N images in a data set is expensive. We generate a list of C exemplars to correspond with each image. By narrowing down from N to C, we decrease the correspondence computation from N 2 times to N C times, which is linear to the number of images in the data set. We generate the list of C exemplars by running PLSA and choosing the top ranked images from each topic. The images are ranked according to P (d|z). In the experiments, we use 10 exemplars per topic. See Figure 3 for examples where an image is matched to some exemplar images.

3.1. Finding Correspondences We want to find the correspondence between patches across two images, di and dj , that are appearance-wise and geometrically consistent. Suppose there are n patches in di . It would be naive to find the single best match based on appearance and would not give a geometrically consistent correspondence. Instead, we use the correspondence algorithm in [13] to find out the appearance and geometric consistent matches. We first find the k-nearest neighbors for each patch based on appearance, with k large enough so that only appearance-wise very disagreeing matches are excluded. Suppose candidate match a marries patch p in di and patch p in dj , written in shorthand as a = p, p . The appearance affinity of the candidate match a, denoted by A(a), is calculated as the correlation coefficient of the feature descriptors of patch p and patch p . Suppose b = q, q  . Then the geometric affinity is defined as: → −→ − (pp − qq  ) ) (1) G(a, b) ≡ max(0, 1 − Cd ·  −→ − → pp qq   G(a, b) is dimensionless so it does not change when the two − → −→ vectors pp and qq  are multiplied by a constant. Cd controls how tolerant this metric is to distortion. The final affinity matrix M has elements M (a, b) = G(a, b)A(a)A(b), where a = p, p  and b = q, q  . Pairs of candidate matches will have low affinity M (a, b) if either the geometric affinity or one of the appearance affinities is low. The correspondence algorithm we adopt from

[13] figures out the final geometrically and appearance-wise consistent matches based on the dominant eigenvector of the affinity matrix M . Partial matching is achieved by setting the parameter Cd in Equation 1 large enough (we use Cd = 1.5), so that candidate matches that potentially match by appearance but distort the correspondence too much are excluded from the final result.

3.2. The reward map The geometrically and appearance-wise consistent correspondences that are found in the previous section tell us which patches often co-occur; it also ensures us that, when patches co-occur, they co-occur in a geometrically consistent manner. We count the number of matches each patch has and create a “reward map” (Figure 1). In the context of unsupervised object detection, we would expect patches from the object of interest to have more matches, thus the reward value is a good indicator of whether a patch belongs to the object of interest. Using the reward map to locate the object of interest is often not good enough. As we will explain later in better detail, the reward map is indeed correlated with the topic variables z, but the correlation is not high enough to provide accurate estimation of the topic. In fact, the“quality” of the reward map depends on the number of exemplars ones uses. Unless we use a very large set of exemplars (which would be inefficient, because finding correspondences is at least linear to the number of exemplars), many patches from the foreground object will not have consistent matches in other images. This is because the intra-class variation of similar objects in the real world is often very large, and having appearance-wise and geometrically consistent matches is rare. Instead of directly using the reward map to locate the object, we use the framework in Section 2.4. We consider the reward value of each patch as an additional feature r that is related to the hidden topic variable z. Empirically, we found that using P (r|z) performs better than using P (r|d, z), probably because of fewer parameters and less overfitting. We learn the parameters using the EM algorithm:

E − step :

P (w|z) = k2



P (z|d) = k3

w,r

m(d, w, r)P (z|d, r, w)

(5)

w,d

where k1 , ..., k4 are normalization constants, and m(d, w, r) is a co-occurrence matrix that keeps the counts of triples (d, w, r) [11]. A typical distribution of P (r|z) is shown in Figure 4. Notice that we have quantized the reward values into 4 bins using k-means quantization. It is interesting to see that background topic zBG has almost all its rewards concentrated at the first bin (because patches originated from background clutter often have very few matches), while the foreground topic zF G has rewards distributed more evenly. Still, the first bin of both topics is highly concentrated, which implies that directly using reward values to tell foreground from background has a lot of ambiguity. This is still true if we quantize the rewards into a larger number of bins. Hence, instead of directly using the reward values for object detection, we integrate this correspondence-based information into a topic model. The EM algorithm will figure out from data how to make judgements from these two “sensors”: the correspondence sensor (which provides rewards, {r}) and the appearance sensor (which provides visual words, {w}). This integration allows us to use a very small number of exemplars. Even so, the performance significantly improves over traditional topic models, A1 to A3, as we will show later. r1

r2

r3

r4

zFG

0.36

0.37

0.20

0.07

zBG

0.97

0.02

0.01

≈0.00

Figure 4. A typical P (r|z).

3.3. Remarks By using correspondences, we introduce spatial configuration into topic models. We don’t make assumptions and postulate a model for the shape of the object. Neither do we make assumptions as in A2 and A3, about the location and clustering of the object. These are all implicitly taken into account by using the reward map. It will be of future interest to combine A2 and A3 together in this correspondencebased topic model framework. Here is a summary of the advantages of this framework: 1. The nice property of PLSA (namely, handling polysemy) is inherent in the new method.

m(d, w, r)P (z|d, r, w) (3)

2. PLSA can only handle polysemy across documents but not within documents: the same visual word wj can be assigned to different topics in different images (context dependency) but a visual word within an image will always be assigned the same topic, regardless of its spa-

d,r





(2)

P (z|d, r, w) = k1 P (z|d)P (w|z)P (r|z) M − step :

P (r|z) = k4

m(d, w, r)P (z|d, r, w)

(4)

(a)

tial relationship with other patches. The additional feature r allows handling polysemy within a document.

Classification Proposed A1 A3 100 50 50

3. In situations where finding geometrically consistent matches is difficult (e.g. , when objects have large deformations), methods that purely rely on correspondence would fail. In this case, the reward map would turn out flat or erroneous. However, our method will learn this fact and rely on the appearance information instead of the reward map. 4. Topic models can take advantage of information from background clutter. Pure correspondence-based methods cannot.

4. Experiments 4.1. Small objects (synthetic data) The task is to detect dumbbells in cluttered scenes, and we gradually increase the amount of background clutter. In Table 1, we see the proposed method allows heavier background clutter than the other methods.

x1 x2 x5 x10

Classification Proposed A1 100 83 88 64 68 52 62 53

A3 88 69 54 51

Localization Proposed A1 86 74 78 62 65 53 56 47

A3 81 66 52 52

Table 1. Classification and localization accuracy (%) for the small objects experiment in Section 4.1.

Localization Proposed A1 96 50

(b)

Size fixed Size varied Distortion Occlusion Combination Multiple

Classification Proposed 100 86 91 84 81 100

from the clutter apart. This is demonstrated in Table 2(a). To demonstrate the effect of occlusion, we removed some patches from the dumbbell. To verify multiple instance detection, two or three distorted versions of the dumbbell were embedded randomly in the scene. These variations are shown in Table 2(b). Note that the number of objects in the scene is unknown beforehand and the same parameters are used throughout these experiments.

4.3. Unsupervised categorization and localization We use the Caltech-4 and the Caltech-background data set to perform binary classification and localization experiments. Patches are detected by the Hessian Affine interest point detector. We use a codebook size of 500 for quantizing the SIFT descriptors into visual words. The SIFT descriptors are then projected from 128 to 30-dimension using Principal Component Analysis.

Motorbike Airplane

4.2. Confusing background (synthetic data) Synthetic images are generated by embedding nonlinear distortions (using pinch, punch, and perspective transforms with the software Paintshop) of the dumbbell in cluttered backgrounds (Figure 5). To confuse the appearance, we insert patches from the object into the background. Clearly, if appearance alone were used to classify the patches, there is no way to distinguish object from clutter, because it is the spatial configuration of the patches that tells the object

Localization Proposed 96 83 86 84 79 92

Table 2. Classification and localization accuracy (%) for Section 4.2. (a) A1 and A3 perform poorly because the data is ambiguous in appearance. (b) The proposed method still works well under different kinds of distortions.

Equal Error Rate

Figure 5. Synthetic images for Section 4.2. Topic models A1 and A3 cannot distinguish between the object and the background because it is the spatial configuration of the patches that tells the object from the background apart. See performance in Table 2.

A3 50

Face Car Rear

Area under ROC

Proposed

A1

A3

Proposed

A1

A3

1.9 3.8 2.0 8.6

12.5 10.2 5.1 18.8

3.3 13.4 2.3 22.3

99.8 99.1 99.5 92.3

91.2 95.7 96.0 88.3

99.5 92.6 98.9 88.0

Table 3. Image-level classification results (%).

Table 3 shows the receiver-operating characteristic (ROC) equal error rates (EER) and the area under ROC curve. Clearly, methods that consider spatial information outperform A1 (PLSA). Figure 6 has further analysis on the motorbike data set. Figure 7 shows the localization performance. Scores are computed based on the posterior probability P (z|d, w) in A1, and based on P (z|d, w, r) in A3 and in Proposed. The proposed method shows significantly better performance in localization. Figure 8 shows some localization results. The top 15 highest scoring patches are indicated by the yellow

Number of images

300 250 200 150 100 50 0 150 300 450 600 750 900 1050 1200 1350 1500

Number of incorrectly classified images

Number of parts per image 50

Proposed A3 A1

40 30 20 10 0 150

300

450

600

750

900 1050 1200 1350 1500

Number of parts per image

Figure 6. Classification of Caltech motorbike versus Caltech background images. The top figure shows that, among the motorbike images, most images have around 300 to 600 patches (foreground plus background). The bottom figure shows the number of wrongly classified motorbike images with respect to the number of patches in the image. Together, we see that the proposed method performs better than A1 and A3. We also see that images with fewer patches are more often classified incorrectly. The proposed method classifies incorrectly 16 out of 826 motorbike images, all of them having less than 150 patches in the image. Caltech Motorbike

1

(a) A1

Detection Rate

0.8 0.6

(b) A3

(c) Proposed

Figure 8. Localization results.

Proposed A3

0.4 A1

0.2 0 0 1

2

3

4

5 6

7

8

9 10

FP per Image

Figure 7. Patch-level classification result on Caltech motorbike data set. For each method, its top 20% confident patches are classified as foreground versus background; the closer the posterior probability P (z|d, r, w) (or P (z|d, w) in A1) of a patch is to 0 or 1, the higher the confidence of the patch.

ellipses. Notice that the proposed method has much less false alarms than A1 (PLSA) and A3. To make the task more challenging, we selected 473 out of the 826 motorbike images that contain background clutter and discarded the rest that have uniform background, and re-run the above experiments. The equal error rate dropped only by 0.2%.

4.4. Enhancing Google’s image ranking We use images returned by Google’s Image Search (collected in [8]) and run our algorithm to enhance the rank-

ing. To combine the classification results of Google’s result and our image categorization result, we use the average rank vector returned by Google’s image search and our algorithm. Using weighted sum of rank vectors has been previously used in image retrieval and web search literature [10]. We rank the images according to P (d|z). In this experiment, we did not run RANSAC as in [8], which should further enhance our performance. This result is obtained without using any separate labeled data. Results in Figure 9 show that “Google+Proposed” can improve the precision by more than 15% at low recall (0.06).

5. Conclusions and Future work We have proposed a novel approach to unsupervised object detection and image categorization. Our method shows how topic models can benefit from finding correspondences and using the “reward map” as an additional feature. We have shown that, traditional topic model approaches can perform poorly when the appearance of the background clutter is extremely confusing. We have also shown that our method is far more robust than traditional topic mod-

1

Google Car Rears Google+Proposed

0.9

Google+A3

Precision

0.8

Google

0.7 0.6 0.5 0.4 0

0.1

0.2 0.3 Recall

0.4

0.5

Figure 9. Re-ranking of Google images.

els when the amount of background clutter is significantly larger than the number of patches from the object of interest, which is a major problem when applying traditional topic models to images. Superior performance on the Caltech and Google data sets also validate the practicalness of the new approach. In our experiments, the number of topics was specified as the number of object classes expected in the data set. If the number of topics in the data set are very high or when it is unnatural to pre-specify the number of topics, tools from Bayesian analysis such as the Dirichlet process mixture models [2] should be considered. Some prior work improve the image representation by using more structured features [1][5] or denser representations [18][12]. Using these representations should further enhance the performance of our approach.

6. Acknowledgements This work is supported by the Taiwan Merit Scholarship TMS-094-1-A-049 and by the ARDA VACE program.

References [1] A. Agarwal and B. Triggs. Hyperfeatures multilevel local coding for visual recognition. In European Conf. on Computer Vision, 2006. [2] C. Antoniak. Mixtures of dirichlet processes with applications to bayesian nonparametric problems. The Annals of Statistics, 2(6):1152–1174, 1974. [3] S. Belongie, J. Malik, and J. Puzicha. Shape matching and object recognition using shape contexts. IEEE Trans. Pattern Analysis and Machine Intelligence, 24:509–522, 2002. [4] A. Berg, T. Berg, and J. Malik. Shape matching and object recognition using low distortion correspondences. In IEEE Conf. Computer Vision and Pattern Recognition, 2005. [5] B. Epshtein and S. Ullman. Identifying semantically equivalent object fragments. In IEEE Conf. Computer Vision and Pattern Recognition, 2005.

[6] L. Fei-Fei and P. Perona. A bayesian hierarchical model for learning natural scene categories. In IEEE Conf. Computer Vision and Pattern Recognition, 2005. [7] P. Felzenszwalb and D. Huttenlocher. Pictorial structures for object recognition. Intl. Journal of Computer Vision, 61(1):55–79, 2005. [8] R. Fergus, L. Fei-Fei, P. Perona, and A. Zisserman. Learning object categories from Google’s image search. In IEEE Intl. Conf. on Computer Vision, 2005. [9] K. Grauman and T. Darrell. Unsupervised learning of categories from sets of partially matching image features. In IEEE Conf. Computer Vision and Pattern Recognition, 2006. [10] T. H. Haveliwala. Topic-sensitive pagerank: A contextsensitive ranking algorithm for web search. IEEE Trans. Knowledge and Data Engineering, 15:784–796, 2003. [11] T. Hofmann. Unsupervised learning by probabilistic latent semantic analysis. Machine Learning, 42:177–196, 2001. [12] F. Jurie and B. Triggs. Creating efficient codebooks for visual recognition. In IEEE Intl. Conf. on Computer Vision, 2005. [13] M. Leordeanu and M. Hebert. A spectral technique for correspondence problems using pairwise constraints. In IEEE Intl. Conf. on Computer Vision, 2005. [14] D. Liu, D. Chen, and T. Chen. Unsupervised image layout extraction. In IEEE Intl. Conf. on Image Processing, 2006. [15] D. Liu and T. Chen. Semantic-shift for unsupervised object detection. In IEEE CVPR Workshop on Beyond Patches, 2006. [16] J. Malik, S. Belongie, J. Shi, and T. Leung. Textons, contours and regions: Cue integration in image segmentation. In IEEE Intl Conf. Computer Vision, 1999. [17] P. Quelhas, F. Monay, J.-M. Odobez, D. Gatica-Perez, T. Tuytelaars, and L. Van Gool. Modeling scenes with local descriptors and latent aspects. In IEEE Intl. Conf. on Computer Vision, 2005. [18] S. Savarese, J. Winn, and A. Criminisi. Discriminative object class models of appearance and shape by correlatons. In IEEE Computer Vision and Pattern Recognition, 2006. [19] J. Sivic, B. Russell, A. Efros, A. Zisserman, and W. Freeman. Discovering objects and their location in images. In IEEE Intl. Conf. on Computer Vision, 2005. [20] M. Weber, M. Welling, and P. Perona. Unsupervised learning of models for recognition. In European Conf. on Computer Vision, 2000.

Unsupervised Image Categorization and Object ...

cal proportions of visual words and have shown promising results. In this paper we will ... Analysis (PLSA)[11], were originally used in the text un- derstanding ...

2MB Sizes 1 Downloads 225 Views

Recommend Documents

Unsupervised Segmentation and Categorization of Skin ...
In the sequence we use stochastic texture fea- tures to refine the suspicious ... 1: Illustration of the three segment categories obtained by the proposed method.

Structural Context for Object Categorization
77 matches - State Key Laboratory for Novel Software Technology,. Nanjing .... ment results of Object Categorization on Caltech-101 are included in section 4.

Unsupervised image layout extraction
ABSTRACT. We propose a novel unsupervised learning algorithm to ex- tract the layout of an image by learning latent object-related aspects. Unlike traditional ...

Semantic-Shift for Unsupervised Object Detection - CiteSeerX
notated images for constructing a supervised image under- standing system. .... the same way as in learning, but now keeping the factors. P(wj|zk) ... sponds to the foreground object as zFG and call it the fore- ..... In European Conference on.

Unsupervised image layout extraction
from the Caltech face dataset [9]. As in [7], images are re- sized to around 200 × 140 and converted to grayscale. A random subset of 50 face and 50 non-face ...

Object Categorization in the Sink: Learning Behavior ...
of Electrical and Computer Engineering, Iowa State University. {shaneg, sukhoy, twegter .... from Lowe's (a home improvement store). The sink fixture contains a ...

Improving Part based Object Detection by Unsupervised, Online ...
based on online boosting to improve the performance on ... based image/video retrieval, etc. Recently, the boosting .... number of sequences from the CAVIAR video corpus [19], and a number of ..... In the Fifth International Conference on.

Unsupervised Learning of Probabilistic Object Models ...
ing weak supervision [1, 2] and use this to train a POM- mask, defined on ... bined model can be used to train POM-edgelets, defined on edgelets, which .... 3. POM-IP and POM-edgelets. The POM-IP is defined on sparse interest points and is identical

Modeling Image Context using Object Centered Grid - Irisa
will form the scene features in a much more coherent and informative way than .... of structuring elements, color distribution and semantic relations. Distribution of ..... network.org/challenges/VOC/voc2007/workshop/index.html. [24] C. C. Chang ...

Image Transformation for Object Tracking in High ...
and a wireless data transmitter walks across the field of view of a camera. In case of a ..... Computer Vision, Cambridge Univ. Press, March 2004. [5] S. Park and ...

Modeling Image Context using Object Centered Grid - CSC - KTH
shown the importance of contextual information in computer vision tasks, like .... An independent. BoF is computed for each grid cell and then all of the BoFs are concatenated ... Of course, one cannot expect the cells of an object- centered grid ...

Modeling Image Context using Object Centered Grid - CSC - KTH
shown the importance of contextual information in computer vision tasks, like ... size of different classes of objects in an image [1]. ... vision as well as the basics of our OCG idea. ..... [Online]. Available: http://lear.inrialpes.fr/pubs/2007/MS

Image segmentation approach for tacking the object: A Survey ... - IJRIT
Identification: Colour is a power description tool, for example, the red apple versus the brown ..... Code available at http://www.cis.upenn.edu/˜jshi/software/.

Image segmentation approach for tacking the object: A Survey ... - IJRIT
Identification: Colour is a power description tool, for example, the red apple versus the ..... Conference on Pattern Recognition., Quebec City, Canada, volume IV, ...

Experiments with Semi-Supervised and Unsupervised Learning
The system of Martin. (1990) explained linguistic metaphors through finding the corresponding metaphorical ...... Source Cluster: polish clean scrape scrub soak.

Automatic Text Categorization
Dec 5, 2006 - novelty presented, is the application of ML based ATC to sentiment classification. The corpus used was collected from the Internet Movie Database (IMDb) archive of the rec.arts.movies.reviews newsgroup. Ratings were extracted and con- v

Vagueness and Order Effects in Color Categorization - David Ripley
Dec 13, 2013 - Springer Science+Business Media Dordrecht 2013. Abstract This paper ... relatively insensitive to small differences: if a predicate like “tall” applies to an individ- ual, then it ..... on the number they were assigned as participa

Vagueness and Order Effects in Color Categorization - David Ripley
Dec 13, 2013 - two color sets (for instance we did not try to ensure that the distance between stimulus. 10 and 11 ... used a constant time lag between each answer and the next stimulus, of 200 ms, with a time lag ..... yellowness in it, even if it i

Implicit and explicit categorization of natural scenes
Abstract: Event-related potential (ERP) studies have consistently found that emotionally arousing (pleas- ant and unpleasant) pictures elicit a larger late positive potential (LPP) than neutral pictures in a window from 400 to 800 ms after picture on

Image retrieval system and image retrieval method
Dec 15, 2005 - face unit to the retrieval processing unit, image data stored in the image information storing unit is retrieved in the retrieval processing unit, and ...

Iterative Learning Control: Brief Survey and Categorization - IEEE Xplore
Index Terms—Categorization, iterative learning control (ILC), literature review. ... introduction to ILC and a technical description of the method- ology. In Section II ...

Unsupervised, Efficient and Semantic Expertise Retrieval
a case-insensitive match of full name or e-mail address [4]. For. CERC, we make use of publicly released ... merical placeholder token. During our experiments we prune V by only retaining the 216 ..... and EX103), where the former is associated with

Unsupervised, Efficient and Semantic Expertise Retrieval
training on NVidia GTX480 and NVidia Tesla K20 GPUs. We only iterate once over the entire training set for each experiment. 5. RESULTS AND DISCUSSION. We start by giving a high-level overview of our experimental re- sults and then address issues of s