ViewFocus: Explore Places of Interests on Google Maps Using Photos with View Direction Filtering Zhiping Luo, Haojie Li, Jinhui Tang, Richang Hong, Tat-Seng Chua School of Computing, National University of Singapore

{luozhipi, lihj, tangjh, hongrc, chuats}@comp.nus.edu.sg

ABSTRACT This paper presents a novel system to explore places of interests based on the large amount of photos that are placed on Google Maps. The system, named ViewFocus, estimates the view directions of photos via robust object matching and camera reconstruction techniques, and geo-registers the directions on the map. Thus users are able to select the places they are interested in, and the system automatically returns a set of precise photos of the target places for users to focus their exploration, by filtering out photos that are pointing to other directions.

Figure 1 shows the example for “Taihe Palace”, a famous building in “forbidden city” in Beijing, China. Figure 1 (a) shows the top 12 most likely photos of “Taihe Palace” selected by 5 users from the photos retrieved from Google Panoramio system [2], where incorrect photos are marked with red boxes. We can see that only 4 out of the 12 photos are correct. While the 12 photos (Figure 1 (b)) found by our system are all correct.

Categories and Subject Descriptors H.5.3 [Information interface and presentation]: Group and organization Interfaces-organization design, Web-based interaction

General Terms Algorithm, Design, Experimentation

Keywords world explore, view direction, map interface

1.

INTRODUCTION

Google Maps [1] is a widely used online service to explore places on the earth. Users can virtually explore the maps by viewing the related geo-tagged photos. Such service provides people an easy way to explore the world with “being there” experience. However, current service mainly relies on the geographical metadata of photos. This has limited its applications. Imaging the following scenario: when you are exploring a place on the high resolution map, you find a region (maybe a building or landscape site) you are interested and want to view its visual appearance by browsing the corresponding photos. One possible way is to zoom the map and let the system return the photos near the desired region. But, this will introduce many erroneous photos. On one hand, photos taken by non location-aware devices may be wrongly placed manually on the map by uploaders. On the other hand, even the photos are correctly placed (manually by users or automatically from the location-aware devices), their viewing directions may not be pointing to the desired region.

Figure 1: (a) the top 12 most likely browsed photos for the place “Taihe Palace”. (b) 12 photos found by our system. Motivated by these observations, we propose a novel place exploration system, named ViewFocus, which takes photo’s view direction into consideration and supports more precise exploration. It works as follows (see Figure 2). Given a desired region, it first collects a set of photos which fall into the range of 100-meter radius centered at the region. Second the photos’ view directions are estimated and registered to the map. Finally only photos that are pointing to the region are returned to the users for exploration.

Copyright is held by the author/owner(s). MM’09, October 19–24, 2009, Beijing, China. ACM 978-1-60558-608-3/09/10.

Figure 2: System pipeline of ViewFocus.

A similar system to ViewFocus is “look around” [2], the new feature released by Google Panoramio recently. “Look around” allows users to watch places from different perspectives by browsing a set of closely matched photos. However, to explore a specified place/region on the map, users need to first know at least one photo shot to the place and start exploring from that photo. On the other hand, ViewFocus, dedicated to focused exploration, can automatically select the photos that point to the desired place for users’ viewing, by filtering nearby photos using recovered view directions.

2.

THE APPROACH

We present the technical details of view direction estimation and registration in selecting photos that are pointing to the target region.

2.1 View Direction Estimation We first extract the SIFT [3] features of the photos. Then for each pair of photos, their SIFT descriptors are matched. If the matches is larger than an empirically determined threshold t, where t = 25 in this work, this pair remains. The reason for choosing highly matched pairs is because they facilitate the reconstruction of the camera parameters. Next, we employ a bundle package [4] to estimate the external camera parameters (Rotation R and Translation T ) of the matched photos. Using a pinhole camera model, the 3D view direction Vp of the photo p can be obtained as follows. ′



Vp = R ∗ [0 0 − 1]

(1)



where indicates the transpose of a matrix or vector. We retain the x and y components of Vp as the 2D view direction vp of p.

create a m × 2 matrix M. Each row of M is an estimated view direction of p. We then apply RANSAC [5] analysis to M to select inliers as the correct view directions. At last, the average of the inliers is computed as the final view direction of p on the map. We repeat the process until all photos’ view directions are geo-registered to the map.

3. DEMONSTRATION INTERFACE The interface of ViewFocus consists of four panels: control panel, map panel, thumbnail panel and slide show panel. Figure 3 shows the screenshot of the interface. In the control panel, the location search with “Explore” button allow users to fly to the desired map. Users can “walk” on the map by dragging it. When users are interested in one region, they can click the “Select” button and draw a rectangle on the map using the left mouse button. By clicking the “All photos” button, the system will return all the photos lying in a range of 100-meter radius centered at the specified region. While clicking “ViewFocus” button will return only the photos point to the region. The filtering performance of the proposed system can be clearly demonstrated by alternately clicking these two buttons. In the thumbnail panel, the photos selected by “ViewFocus” or “All photos” are organized as a photo list, allowing quick reviewing of the results by the users. In the slide show panel, interactive photo zooming and panning are supported. When users single-click a photo thumbnail, its high-resolution image will be shown in this panel. And double-clicking a thumbnail will result in a slide show of the photo list starting from the clicked photo. When a photo is being shown, its location on the map is indicated with a red mark.

4. CONCLUSION This paper presented a system called ViewFocus. It supports focused exploration of places on the map with relevant photos whose view directions are pointing towards the regions. As more and more geo-tagged photos are becoming available [6], we believe that ViewFocus provides an intuitive and desirable way to explore the world.

5. REFERENCES

Figure 3: User interface of ViewFocus.

2.2 View Direction Geo-registration Now we have the relative camera locations and their view directions in a world coordinate system. The next step is to align the view directions to the map. The geo-registration of view direction vp of photo p is carried out as follows. Let q be one of the matched photos of p. Because the angles determined by p′ s view direction and the location vector from p to q are unchanged in the world coordinate system and map coordinate system, the view direction vpq of p in the map based on q can be obtained by using a rotation transformation. Suppose there is a total of m photos matched with p, we compute m view directions for p from these photos and

[1] Google Maps, http://maps.google.com/ [2] Panoramio, http://www.panoramio.com/ [3] D.G. Lowe, “Distinctive Image Features from Scale-Invariant Keypoints”, Int. J. Comput. Vision, 60(2), pp. 91-100, 2004 [4] M.I.A Lourakis and A.A. Argyros, “SBA: A software package for generic sparse bundle adjustment”, ACM Trans. Math. Softw., 36(1), pp. 1-3, 2009 [5] M.A. Fischler and R.C. Bolles, “Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography”, Comm. of ACM, 24(6), pp. 381-395, 1981 [6] Y.-T Zheng and M. Z and Y. S and H. A and Y. B and A. B and F. Band T.-S Chua and H. N, “Tour the World: building a web-scale landmark recognition engine”, CVPR’09, 2009

ViewFocus: Explore Places of Interests on Google ...

Google Maps [1] is a widely used online service to explore places on the earth. Users can virtually explore the maps by viewing the related geo-tagged photos.

205KB Sizes 1 Downloads 195 Views

Recommend Documents

Detecting Communities with Common Interests on Twitter
Jun 28, 2012 - Twitter, Social Networks, Community Detection, Graph Mining. 1. INTRODUCTION ... category, we selected the six most popular celebrities based on their number of ... 10. 12. 14. 16. 18. Control Group. Film & TVMusic Hosting News Bloggin

Recommending Tours and Places-of-Interest based on ...
*Department of Computing and Information Systems, The University of Melbourne, Australia. †Victoria ... travel sequences of a user based on his/her geo-tagged.

Statement of research interests - Etienne Laliberté
May 22, 2009 - I have also recently conducted a meta-analysis on the impacts of land use ... I have recently developed the FD R package (http://cran.r-.

Inferring Implicit Topical Interests on Twitter
their social posts, less work is dedicated to identifying implicit interests, which are also very important for building .... dedicated to inferring implicit interests of the users. Some authors have shown interest in ... In Definition 2, For instanc

Explore and Challenge - GitHub
Open Scratch GPIO from the desktop using the Scratch GPIO icon (we do not need the ... This is where you build your scripts by locking various blocks together.

Personalized Tour Recommendation based on User Interests and ...
terest preferences of users or adhere to their trip constraints. Furthermore, the massive volume of information ... personalized tours with POIs and visit duration based on POI popularity, users' interest preferences and ...... SAMAP: An user-oriente

Explore Transformations of Rational Functions.pdf
Page 1 of 3. Explore Transformations of Rational Functions! Go to: https://www.desmos.com/calculator. 1. Graph. a. As x gets larger (approaches infinity) what ...

Explore and Challenge - GitHub
Select the Variables tab and add a new variable by pressing the "Make a variable" button, call it Score and set it to be For all sprites. We will also need to create a list to hold our sequence of lights, we will call it GameList: Press the "Make a l

Explore and Challenge - GitHub
Explore and Challenge Scratch GPIO: Pi-Stop Traffic Sequence - Create your own ... Once you have started the Raspberry Pi desktop, open Scratch using the ...

Explore and Challenge - GitHub
WORKSHEET: Tick the checkbox marked "I've created the Pi-Stop STOP and GO sequences". The Final Program - Changing Lights. At the moment our program ...

Statement of research interests - Etienne Laliberté
May 22, 2009 - domain have been to clarify and develop multivariate methods for analyzing spatial patterns and quantifying the importance of niche and other ...

Statement of research interests - Etienne Laliberté
May 22, 2009 - Empirical ecology has made enormous recent progress in evaluating the impacts of human ... A next frontier is to define interactions not by the.

Google Places
Click “Add New Business” to claim your listing ... your listing today with Google Places, to attract customers and help your business stand out from competitors. At.

Places API
The Google Places API for Android can provide a great Autocomplete experience for address entry in your mobile application. Places API. Improved customer ...

Explore 18.1.5 ANSWERS.pdf
Download. Connect more apps... Try one of the apps below to open or edit this item. Explore 18.1.5 ANSWERS.pdf. Explore 18.1.5 ANSWERS.pdf. Open. Extract.