2013 13th International Conference on Control, Automation and Systems (ICCAS 2013) Oct. 20-23, 2013 in Kimdaejung Convention Center, Gwangju, Korea

Evaluation of Vocabulary Trees for Localization in Robot Applications Soonmin Hwang, Chaehoon Park, Yukyung Choi, Donggeun Yoo and In So Kweon∗ ∗

Department of Electrical Engineering, KAIST, Daejeon 305-701, Korea Corresponding author (Tel : +82-42-350-5465; E-mail: [email protected])

Abstract: Vocabulary tree based place recognition is widely used in topological localization and its various applications have been proposed during the past decade. However, the bag-of-words representations from the vocabulary tree, which is trained with fixed training data, are difficult to be optimized to dynamic environments. To solve this problem, an adaptive vocabulary tree has been proposed, but there has been no comparison considering the adaptive properties of the conventional vocabulary tree. This paper provides a performance evaluation of the vocabulary tree and the adaptive vocabulary tree in dynamic scenes. This analysis provides guidance for choosing appropriate vocabulary in robot applications. Keywords: Topological localization, Image Retrieval, Vocabulary Tree

1. INTRODUCTION As the robot industry is developing rapidly, the demands for robots are also increasing. These robots need to know the current position for doing many tasks, for example, indoor service. However, it is difficult for the robot to use a GPS sensor to recognize its own location, because the robot usually works in an indoor environment. As the GPS sensor is unhelpful in indoor situations, we can use visual information to recognize position. This problem is called ”visual-SLAM” (Simultaneous Localization and Mapping). One of the most significant requirements for visual SLAM is robust place recognition [1]. In recent years, several algorithms for comparing images as numerical vectors in the bag-of-words model have been introduced [2]. This model results in very effective and quick image matches [3], [4]. The vocabulary tree (VT) [3] is a simple and powerful method for image indexing and retrieval, but many methods based on the bag-of-words model needs an offline training step. It is built with prepared training images. This kind of pre-processing is inefficient and requires prior knowledge about the given environment. The static representations from the vocabulary tree, which is trained with fixed training data, are difficult to be optimized to dynamic environments. To overcome these problems, an adaptive vocabulary tree (AVT) approach [4] has been proposed. The adaptive vocabulary tree incrementally and continuously adapts vocabularies with incoming database images. In [4], T. Yeh et al. simply showed that the performance of the AVT declines as the vocabulary stops adapting, so they concluded that the conventional VT cannot be used with incoming images. Even though the AVT is robust against dynamic environments, for updating the vocabulary tree, it needs additional memory for maintaining training data and the computational burden. The aim of this work is to provide an evaluation of the adaptive properties of the dynamic data and guidance for choosing between the VT and the AVT according to applications. We evaluate the standard VT and the AVT in

terms of the building time, indexing time, storage size and recognition performance in dynamic environments. The remainder of this paper is organized as follows: preliminary information is presented in section 2, the framework for evaluation is explained in section 3, then in section 4 we discuss some experimental results, and conclude this paper in section 5.

2. PRELIMINARY Bag-of-words model In the field of image recognition, the most popular method is the bag-of-words model which was proposed in 2003 by Sivic et al [2]. In this model, some representative features are defined by clustering whole features in the training images. These are called visual words. Then an image can be represented as a histogram vector indicating the number of visual words in the image. A set of visual words is called a codebook or visual vocabulary. The performance of retrieval depends heavily on the distinctiveness of the vocabulary. Distinctiveness of vocabulary The simplest way to make a more distinctive vocabulary is to make a large vocabulary. As more visual words are obtained, a cell represented by a visual word becomes smaller. Then the loss of information caused by quantization is reduced. In other words, an image can be represented using more visual words. Thus, some images can be distinguished by a large vocabulary (Figure 1). That is to say, a histogram vector using more distinctive vocabulary might be more distinguishable. Thus, the distinctiveness of vocabulary is estimated by retrieval accuracy. However, a flat vocabulary, which is made by conventional k-means clustering, requires more time to make a large vocabulary and to search corresponding visual words using a large vocabulary. Thus, it is difficult to make a flat vocabulary large. Hierarchical Vocabulary Tree In order to make a large vocabulary efficiently and to use it, D. Nister proposed a hierarchical vocabulary tree[3]. In [3], clusterings to get a small number of visual words compared to the number

(a)

(b)

Fig. 1 (a) Less distinctive vocabulary (b) More distinctive vocabulary. Using more distinctive vocabulary, two images could be distinguishable.

of whole visual words are performed recursively, thereby a hierarchical tree is obtained as a vocabulary. In this way, more visual words are made quickly and it requires a much shorter time to represent an image as a bag-ofwords histogram compared to a linear search. Adaptive Vocabulary Tree According to Yeh et al. [4], a fixed vocabulary model is inadequate in a dynamic environment. They argued that adapting to new data is helpful for improving performance in a dynamic environment. In their approach, a dynamic environment means that object categories or scenes are not yet determined. Adapting is a re-clustering process when there are many features at the leaf node. Basic Image Retrieval Framework The image retrieval framework we used follows the bag-of-words model. This means that we use some kind of vocabulary to represent an image as a numerical vector. The retrieval framework consists of three steps, learning a vocabulary, creating a database, and retrieval. First, the vocabulary is learned by clustering local features. Then dataset images are represented as histogram vectors using the learned vocabulary. Finally, in the retrieval step, a query image is also represented as a histogram vector and the closest vector in the dataset vectors is found using some kind of distance. For speed-up of the retrieval time, an inverted file structure is used. [3]

3. EVALUATION FRAMEWORK We use a SIFT descriptor [5] which was proposed by D. Lowe and is the most popular descriptor. To compare the distinctiveness of vocabularies, we employ the image retrieval framework which is a solution of the topological localization of a robot. We also check the times needed to build a vocabulary and to index images, which means a query image is represented by a vector. In addition, we consider memory occupancy. We use ukbench[3] for the dataset and oxford5k[7] for vocabulary learning. These are the most popular datasets in image retrieval. The ukbench consists of 10,200 images such as book covers, pictures, and household items. The images include 2,550 objects and undergo some

transformation such as rotation and scaling. Thus, each object appears 4 times in the ukbench dataset. Considering the scale of the robot application, we select 1,000 images for our experiment. The oxford5k consists of 5062 images including particular oxford landmarks. We use a subset of oxford5k for the learning vocabulary. In our experiments, 1,000 images in ukbench are utilized as a database and the same images are also used for queries. For learning a vocabulary, static and dynamic environments are considered. The static environment is a situation in which we already know what images should be recognized, so that we build a vocabulary using target images. In this case all features in the dataset are already known, so the most representative visual words can be made. On the other hand, a dynamic environment means that the task is not yet determined, so we use some images not included in the database images to build a vocabulary. In the case of the AVT, the first step is omitted. The AVT is learned simultaneously using database images in the second step. We evaluate the adaptiveness of a static vocabulary in a dynamic environment. In addition to the distinctiveness of vocabulary, memory occupancy and computational time are also important due to a limitation of resources in robot applications. Thus, we evaluate two vocabularies on the basis of retrieval accuracy, the computational time and memory occupancy. Parameters of AVT Figure 6 represents the performance of AVT at various parameters. For fair comparison, 100 for C and 0.5 for S are selected to the best parameters. Parameters of VT We follow general parameter of vocabulary tree, 10 for branch factor and 6 for level.

4. EXPERIMENTAL RESULTS This section presents the results of experiments that compare the vocabulary tree (VT) proposed by Nister[3] with the adaptive vocabulary tree (AVT) proposed by Yeh[4]. In general, a large vocabulary that contains many visual words makes image vectors more discriminative. For a fair comparison, we fixed size of the vocabulary trees. First, we experimented with the recognition rate indicating the distinctiveness of the vocabulary. Next, to determine a more efficient vocabulary, we analyzed the storage size with respect to the recognition rate. Finally, we compared the computation time over the size of the vocabulary. Because there is a trade-off between the size of the vocabulary and the computation time, it is important to analyze the computation time. In our experiments, we considered the static environment and the dynamic environment. Our experiments provide guidance for a more appropriate vocabulary in robot applications. 4.1 Recognition Rate As shown in Figure 2, when the size of one vocabulary was similar to the size of another vocabulary, the

6

2.5

0.75

0.7 VT-UKBench (D) VT-UKBench (I) VT-Oxford5k AVT

0.65

0

0.5

1 Vocabulary Size

1.5

Offline Building time (ms)

Recognition Rate

0.8

2 1.5 1

VT-UKBench (D) VT-UKBench (I) VT-Oxford5k

0.5 0 0

2 x 10

x 10

0.5

5

Fig. 2 Static environment (black and red), Dynamic environment (blue, green and red)

1 Vocabulary Size

1.5

2 x 10

5

Fig. 4 Comparison with building time

6

2

x 10

Storage Size

1.5

1

VT-UKBench (D) VT-UKBench (I) VT-Oxford5k AVT

0.5

0 0

0.5

1 Vocabulary Size

1.5

2 x 10

Indexing Time (ms)

200

150 VT-UKBench (D) VT-UKBench (I) VT-Oxford5k AVT

100

50

0 0

0.5

1 Vocabulary Size

1.5

2 x 10

5

Fig. 5 Comparison with indexing time

5

Fig. 3 Static environment (black and red), Dynamic environment (blue, green and red)

VT was more distinctive than the AVT in a static environment. The VT was made using all the features in all the images, thereby the more representative features became visual words. In this way, the VT achieves a better performance. However, if the size of the AVT increased, it also had good accuracy. In a dynamic environment, when the size of the vocabulary was less than 100,000, the VT, which was learned using unseen images such as in Oxford5k, was a little bit more accurate than the AVT. However, if the vocabulary was large enough, the AVT was better, because, the vocabularies learned by unseen images might include many unnecessary visual words. In this case, even if the vocabularies were large, accuracy remained the same. On the other hand, the AVT continued to adapt to new data. For this reason, the larger AVT had more accurate results. As Nowak mentioned regarding a vocabulary made by unseen images [6], when we used Oxford5k images to build a vocabulary, the recognition rate was not significantly different from others. That means that the VT, which is made by sufficient features, could be distinctive enough to use in a dynamic environment.

4.2 Storage Size Storage size is the sum of memories for a vocabulary tree and additional information. The VT does not need additional information, but the AVT needs to keep all the input features for re-clustering, which is called adapting, so that the AVT is bigger than the VT. (Figure 3) 4.3 Computation Time One of the differences between the VT and the AVT is computation time. The VT needs an offline building process. (Figure 4) On the other hand, the AVT was designed to adapt to new data, so the AVT does not need any pre-processing. This might be one of the advantages of the AVT. Instead, it has another burden: the online processing time. As shown in Figure 5, because the AVT continuously adapts to new data, it takes more time to represent an image as a vector. Therefore, it is an issue: whether we spend time offline or online.

5. CONCLUSION There are some methods for building a good vocabulary tree. To deal with a dynamic environment, the AVT was proposed. However, more concrete analysis is needed for an environment with limited resources, such as a robot application. In this paper, we evaluated the VT and the AVT in terms of distinctiveness, memory and time efficiency in static and dynamic environments. The

100

250 200 150 AVT AVT AVT AVT AVT AVT AVT

100 50 1

2 3 Vocabulary Size

60 40

0 0

5

1

5

350

120

300

100

250 200 150 AVT AVT AVT AVT AVT AVT AVT

100 50 1

2 3 Vocabulary Size

S S S S S S S

=0.1 =0.3 =0.5 =1 =2 =5 =10

4

2 3 Vocabulary Size

4

0.7

0.5 0

5 x 10

1

5

2 3 Vocabulary Size

4

5 x 10

5

(c) AVT AVT AVT AVT AVT AVT AVT

80

S S S S S S S

1

=0.1 =0.3 =0.5 =1 =2 =5 =10

60 40

AVT AVT AVT AVT AVT AVT AVT

0.9

S S S S S S S

=0.1 =0.3 =0.5 =1 =2 =5 =10

0.8 0.7 0.6

20 5

x 10

C =20 C =50 C =100 C =200 C =500 C =1000 C =2000

0.8

(b)

Retrieval Time (ms)

Indexing Time (ms)

(a)

0 0

AVT AVT AVT AVT AVT AVT AVT

0.9

0.6

20

4 x 10

80

1

C =20 C =50 C =100 C =200 C =500 C =1000 C =2000

Recognition Rate

0 0

C =20 C =50 C =100 C =200 C =500 C =1000 C =2000

AVT AVT AVT AVT AVT AVT AVT

Recognition Rate

120

300 Retrieval Time (ms)

Indexing Time (ms)

350

0 0

1

5

(d)

2 3 Vocabulary Size

4

5 x 10

(e)

5

0.5 0

1

2 3 Vocabulary Size

4

5 x 10

5

(f)

Fig. 6 Parameters of AVT AVT copes with a dynamic environment by adapting to new data, so that it does not need any batch process and it becomes a more distinctive vocabulary. However, to adapt to new data, the AVT must hold all the features and be re-clustered continuously. Thus, it needs more memory and has an online burden. Furthermore, according to the results of Oxford5k experiments, the VT made by sufficient features is distinctive enough. In conclusion, if batch processing is not available in a dynamic environment or accuracy may have a high priority, the AVT seems to be more appropriate. If having limited resources is a significant issue, the VT seems to be better.

6. ACKNOWLEDGEMENT This work was supported by National Strategic R&D Program for Industrial Technology, Korea.

REFERENCES [1]

[2]

[3]

[4]

A.J. Glover, “FAB-MAP + RatSLAM: Appearancebased SLAM for multiple times of day,” IEEE International Conference on Robotics and Automation, 2010. J.Sivic and A.Zisserman, “Video Google: A Text Retrieval Approach to Object Matching in Videos,” IEEE International Conference on Computer Vision, 2003. D.Nister and H.Stewenius, “Scalable Recognition with a Vocabulary Tree,” IEEE Conference on Computer Vision and Pattern Recognition, 2006. T.Yeh, J.Lee and T.Darrell, “Adaptive Vocabulary Forests for Dynamic Indexing and Category Learning,” IEEE International Conference on Computer Vision, 2007.

[5]

[6]

[7]

David G. Lowe, “Distinctive Image Features from Scale-Invariant Keypoints,” International Journal of Computer Vision, 2004. E.Nowak, F.Jurie and B.Triggs, “Sampling strategies for bag-of-features image classification,” European Conference on Computer Vision, 2006. J.Philbin, R.Arandjelovic and A.Zisserman, “The Oxford Buildings Dataset,” http://www.robots.ox.ac.uk/ vgg/data/oxbuildings/.

Evaluation of Vocabulary Trees for Localization in ...

to build a vocabulary and to index images, which means a query image is represented by a vector. In addition, we consider memory occupancy. We use ukbench[3] for the dataset and oxford5k[7] for vocabulary learning. These are the most popular datasets in image retrieval. The ukbench consists of 10,200 im- ages such ...

398KB Sizes 1 Downloads 200 Views

Recommend Documents

Dead Reckoning for Monte Carlo Localization in Low ...
Especially in scenar- ios with low seed densities, we find SA-MCL to significantly outperform MCL. S. S. S. S. S. Figure 6: Setup of demo with seeds and RC cars.

A Scalable UWB Based Scheme for Localization in ...
However simple GPS based schemes do not work well ... The goal is to track the vector hn, estimate the channel taps, ..... location and tracking system”, Proc.

Robust Eye Localization for Lip Reading in Mobile ...
Emails: {[email protected], [email protected]}. Abstract- ... setting a proper region including lip. ... detection is to set a sufficient lip region, where the fine lip.

HISTOCHEMICAL-LOCALIZATION-OF-HYALURONATE....pdf ...
intercellular spaces from basal to upper spinous layers displayed strong staining, most intense in the middle. spinous cell layer. The uppermost vital cell layers as well as the cornified cell layer remained unstained. In the non-keratinized epitheli

Evaluation of Writeprints Technique for Authorship Recognition in ...
authorship recognition method, WritePrint, to evaluate it against two simple ad- versarial attacks- obfuscation attack .... Frequency of Parts of speech tag. Syntactic. Punctuation. Frequency and ... (c) Compute window instance coordinates (principal

Blockers for Non-Crossing Spanning Trees in Complete ...
Aug 16, 2011 - T (i.e., the graph obtained from T by removing all leaves and their incident leaf edges) is a path (or is empty). A longest path in a caterpillar T is called a spine of T. (Note that any edge of T either belongs to every spine or is a

pdf-1419\localization-of-clinical-syndromes-in-neuropsychology-and ...
... apps below to open or edit this item. pdf-1419\localization-of-clinical-syndromes-in-neurop ... uroscience-by-joseph-m-tonkonogy-antonio-e-puente.pdf.

Activitydependent localization in spines of the ... - Wiley Online Library
removed, frozen in dry ice and cut coronally with a cryostat microtome into 10- to 30-μm sections. These were fixed in. 4% PFA in 0.1 M PB for 30 min. The subsequent histochemi- cal procedures were as described for Fig. 3A. When F-actin was to be de

External Localization System for Mobile Robotics - GitHub
... the most known external localization reference is GPS; however, it ... robots [8], [9], [10], [11]. .... segments, their area ratio, and a more complex circularity .... The user just places ..... localization,” in IEEE Workshop on Advanced Robo

The Insights of DV-based Localization Algorithms in the ... - CiteSeerX
Email: yuanfang [email protected], Li [email protected]. † ... tion service system. However, it is ... use the distance vector (DV) routing method to broadcast.

Robot Localization Sensor for Development of Wireless ... - IEEE Xplore
issues for mobile robot. We describe a novel localization sensor suite for the development of a wireless location sensing network. The sensor suite comprises ...

Evaluation of Business Solutions in Manufacturing Enterprises
Department of Information and Communication Technology,. University of ... He received his degree in Informatics Engineering from Institut Teknologi · Bandung ...

Robot Localization Network for Development of ...
suite for the development of a location sensing network. The sensor suite comprises ... mobile robotics, localization technology refers to a systematic approach to ...

Fast maximum likelihood algorithm for localization of ...
Feb 1, 2012 - 1Kellogg Honors College and Department of Mathematics and Statistics, .... through the degree of defocus. .... (Color online) Localization precision (standard devia- ... nia State University Program for Education and Research.

The Insights of DV-based Localization Algorithms in the Wireless ...
School of Software, Dalian University of Technology, China. Email: yuanfang [email protected] ... applications in wireless sensor networks (WSNs). However, most.

Selection of Nest Trees by Cavitynesting Birds in the Neotropical ...
6Corresponding author; e-mail: [email protected]. BIOTROPICA ...... Uno de los centros de biodiversidad más importantes para la conservación a nivel.

A stand-alone method for anatomical localization of ...
implemented as part of our NIRS Analysis Package (NAP), a public domain Matlab ... MNI space (Okamoto et al., 2004; Okamoto and Dan, 2005; Singh et al.,.

Convolutional Networks for Localization Yunus Emre - GitHub
1 Introduction: Image recognition has gained a lot of interest more recently which is driven by the demand for more sophisticated algorithms and advances in processing capacity of the computation devices. These algorithms have been integrated in our

Evaluation of Business Solutions in Manufacturing Enterprises
degree from Computer Science Institute of University of Ancona (Italy) in ... The last years have seen the emergence of risk as a metric for prioritizing events ... model, manufacturing SMEs can be distinguished into two main categories: product-.

Relative-Absolute Information for Simultaneous Localization and ...
That is why it is always required to handle the localization and mapping. “simultaneously.” In this paper, we combine different kinds of metric. SLAM techniques to form a new approach called. RASLAM (Relative-Absolute SLAM). The experiment result

Longest common extensions in trees
Jul 9, 2015 - We start with a simple O(1) query O(n logn) space solution. For every ... store the longest common prefix between any two paths adjacent in.

Selection of Nest Trees by Cavitynesting Birds in the Neotropical ...
decay class, species, growth rate, and crown class of trees (Carey. 1983, Whitford 2002 ... pole or carried up to the cavity using single-rope climbing tech- niques. ..... Conservar La Argentina Grant from Aves Argentinas/BirdLife In- ternational ...

Study on efficiency of sawing coniferous trees in ...
very large part of Kosovo with sawn timber (coniferous). The trunks of coniferous species in this region are: pine, fir and spruce which mainly are imported from Montenegro. In this study, we tried to bring interesting data on the subjects of this re