Matching and Ranking with Hidden Topics towards Online Contextual Advertising Dieu-Thu Le, Cam-Tu Nguyen, Quang-Thuy Ha Coltech, Vietnam National University 144 Xuan Thuy, Cau Giay, Hanoi, Vietnam {dieuthu, ncamtu, thuyhq}@vnu.edu.vn Abstract In online contextual advertising, ad messages (ads) are displayed related to the content of target Web pages. It leads to the problem in IR and Computational Advertising: how to select the most relevant ads given the content of a Web page. To deal with this problem, we propose a framework that takes advantage of knowledge from large scale external datasets. Technically, the framework provides a mechanism to discover the semantic relations between Web pages and ads by analyzing topics for them. This helps overcome the problem of mismatch due to unimportant words and the difference in vocabulary between Web pages and ads. The framework has been evaluated through a number of experiments. It shows a significant improvement in accuracy over word/lexicon-based matching and ranking methods.

1. Introduction Since its birth more than a decade ago, online advertising has been growing quickly and become more diverse in both its appearance and the way it attracts Web users’ attention. According to the Interactive Advertising Bureau (IAB), Internet advertising revenues reached $5.8 billion for the first quarter of 2008, 18.2% increase over the same period in 2007. And its growth is expected to continue as consumers spend more and more time online. Studies have shown that the relevance between target Web pages and ads is an important factor to attract customers [3, 4]. In contextual advertising, ads are delivered based on the content of the Web pages that users are surfing. It can therefore provide Internet users with information they might be interested in and allow advertisers to reach their target customers in a non-intrusive way. In order to suggest the “right” ads, contextual matching and ranking techniques are needed to be applied. Different from sponsored search, in which ads are chosen depending on only the keywords provided by users, contextual ad placement depends on the whole content of a Web page. Keywords given by users are often condensed

Xuan-Hieu Phan, Susumu Horiguchi GSIS, Tohoku University Aobayama 6-3-09, Sendai, 980-8579, Japan {hieuxuan, susumu}@ecei.tohoku.ac.jp

and reveal directly the content of the users’ concerns, which makes it easier to understand. Analyzing Web pages to capture the relevance is a more complicated task. Firstly, as words can have multiple meanings and some words in the target page are not important, they can lead to mismatch in lexicon-based matching method. Moreover, a target page and an ad can still be a good match even when they share no common words or terms. To deal with these problems, we present a framework that can produce high quality match by taking advantage of hidden topics analyzed from large-scale external datasets. In particular, the main advantages of the framework are threefold: • Discovering the semantic relatedness: the framework provides a mechanism to analyze implicit or hidden topics for both Web pages and ads, which helps capture their semantic relations. If a Web page and an ad share more common topics, it is likely that they are more relevant. • Reducing data sparseness: as Web pages and ads are written by different people, there is often a difference between their vocabulary. This framework can be able to deal with this problem by expanding both Web pages and ads with their most relevant topics. The added topics in each Web page and ad help reduce the sparseness and make the data more topic-focused. • Expanding the coverage and enhancing the predictability: hidden topics discovered from large-scale external data collections involve many external terms that might not appear in ads or target Web pages. This helps deal with a wide range of Web pages and ads, as well as process future data (i.e. previously unseen Web pages and ads) better. Moreover, our framework is also easy to implement and general enough to be applied in different domains and languages. Through a number of experiments, it indicates that this framework can suggest appropriate ads for contextual ad placement and can be highly practical in reality. The rest of the paper is organized as follows. Section 2 reviews related studies on contextual advertising. Section 3 proposes the framework of matching & ranking with hidden topics. Section 4 describes hidden topic analysis of a

large dataset. Section 5 describes how we infer hidden topics for ads and Web pages as well as how they are matched & ranked with those topics. Section 6 presents careful experiments on contextual advertising with Vietnamese Web. Finally, some important conclusions are given in Section 7.

pLSA [7] and LDA [1], can be useful in many applications. An example is using LDA to build classifiers that deal with short & sparse data [10]. Our framework is also based on hidden topics from a large-scale dataset to capture semantic relations between Web pages and ads.

2. Related Work

3. Page-Ad Matching and Ranking Framework

The success of sponsored search in Web advertising has motivated IR researchers to study content match in contextual advertising. Thus, one of the earliest studies in this area was originated from the idea of extracting keywords from Web pages. Those representative keywords will then be matched with advertisements [12]. However, the relevance of chosen ads based on extracted keywords in this system has not been proved through experiments yet. While extracting keywords from Web pages in order to compute the similarity with ads is still controversial, Andrei Broder et al. [2] proposed a framework for matching ads based on both semantic and syntactic features. For semantic features, they classified both Web pages and ads into the same large taxonomy with 6000 nodes. Each node contains a set of queries. For syntactic features, they used the TF-IDF score and section score (title, body or bid phrase section) for each term of Web pages or ads. Our framework also tries to discover the semantic relations of Web pages and ads, but instead of using a classifier with a large taxonomy, we use hidden topics discovered automatically from an external dataset. It does not require any language-specific resources, but simply takes advantage of a large collection of data, which can be easily gathered on the Internet. A particular challenge in contextual matching is the difference between the vocabularies of Web pages and ads. Ribeiro-Neto et al. [11] focuses on solving this problem by using additional pages. It is similar to ours in the idea of expanding Web pages with external terms to decrease the distinction between their vocabularies. However, they determine added terms from other similar pages by means of a Bayesian model. Those extended terms can be appeared in ad’s keywords and potentially improve the overall performance of the framework. Their experiments have proved that when decreasing the vocabulary distinction between Web pages and ads, we can find better ads for a target page. Following the former study [11], Lacerda et al. [8] tried to improve the ranking function based on Genetic Programming. Given the importance of different features, such as term and document frequencies, document length and collection’s size, they use machine learning to produce a matching function to optimize the relevance between the target page and ads. It was represented as a tree composed of operators and logarithm as nodes and features as leaves. They used a set of data for training and a set for evaluating from the same data set used in [11] and recorded a better gain over the best method described in [11] of 61.7%. Hidden topics analysis using latent topic models, such as

Keyword bid information

Universal Dataset Web pages

Web pages

Model Estimation

Contextual Matching

Ranking

Topic inference adad ad adad ad ad adad ad ad

Advertising messages

Estimated Topic Model Topic inference

ad adad ad ad

ad ad ad adad ad

Advertising messages

Figure 1. A General Framework for Page-Ad Matching & Ranking with Hidden Topics (a) Choosing an appropriate “universal dataset” (b) Doing topic analysis for the universal dataset (c) Doing topic inference for Web pages and ads (d) Page-Ad Matching and Ranking In this section, we present our general framework for contextual page-ad matching and ranking with hidden topics discovered from external large-scale data collections. Given a set of n target Web pages P = {p1 , p2 , . . . , pn }, and a set of m ad messages (ads) A = {a1 , a2 , . . . , am }. For each Web page pi , we need to find a corresponding ranking list of ads: Ai = {ai1 , ai2 , . . . , aim }, i ∈ 1..n such that more relevant ads will be placed higher in the list. These ads are ranked based on their relevance to the target page and the keyword bid information. However, in the scope of our work, we only consider linguistic relevance and assume that all ads have the same priority, i.e. the same bid amount. As depicted in Figure 1, the first important thing to consider in this framework is collecting an external large-scale document collection (a) which is called Universal Dataset. To take the best advantage of it, we need to find an appropriate Universal Dataset for our Web pages and ads. First, it must be large enough to cover words, topics, and concepts in the domains of Web pages and ads. Second, the vocabularies of the Universal Dataset must be consistent with that of Web pages and ads, so that it will make sure topics analyzed from this data can overcome the vocabulary impedance of Web pages and ads. The Universal Dataset should also be preprocessed including noise and stop words removal before analysis to get better results. In general, we can apply any topic models, such as pLSI [7] or LDA [1], to analyze the Universal Dataset. The result of the step (b) is an estimated topic model that includes hidden topics discovered from the dataset and the distribu-

tions of topics over terms. Steps (a) and (b) will be presented more details in section 4. After the estimating process (b), we can again do topic inference for both Web pages and ads based on this model to discover their meanings and topic focus (c). This information will be integrated into the corresponding Web pages or ads for matching and ranking (d). Both steps (c) and (d) will be discussed more in section 5.

4. Hidden Topic Analysis of Universal Dataset Our framework is largely inspired by the recent success of topic modeling research in Machine Learning and NLP. Topic models [1, 5, 7] allow discovering semantic information from data based on the idea that each document is a probability distribution over topics and each topic, in turn, is a mixture distribution over words/terms. In this section, we give a short introduction about latent Dirichlet allocation (LDA) and the use of this model for analyzing a large news document collection to serve as Universal Dataset.

4.1. Latent Dirichlet Allocation (LDA)

z m,n r β

within each document

r ϑm

r ϕk

wm, n

k ∈ [1, K ]

n ∈ [1, N m ]

for the whole data collection

r α

m ∈ [1, M ]

Figure 2. Graphical structure of LDA LDA is a generative graphical model introduced by Blei et al. [1]. Its graphical structure is shown in Figure 2. And the interpretation of its probabilistic generation process is as → in Table 1. First, a document − w m = {wm,n } (n = 1..Nm ) − → is generated by picking a distribution over topics ϑ m from → a Dirichlet distribution (Dir(− α )), which determines topic assignment for words in that document. Then the topic assignment for each word placeholder [m, n] is performed by sampling a particular topic zm,n from multinomial distri− → bution M ult( ϑ m ). And finally, a particular word wm,n is generated for the word placeholder [m, n] by sampling from → multinomial distribution M ult(− ϕ zm,n ). To estimate LDA, some approximate methods can be used, such as Variational Methods [1] and Gibbs Sampling [5]. Gibbs Sampling is a special case of Markov-chain Monte Carlo (MCMC) and often yields relatively simple algorithms for approximate inference in high-dimensional models such as LDA [6]. Here, we only show the most important formula that is used for topic sampling for words. → → Let − w and − z be the vectors of all words and their topic assignment of the whole data collection W . The topic assignment for a word depends on the current topic assignment of

Table 1. Generative process for LDA for all topics k ∈ [1, K] do − → → sample mixture components − ϕ k ∼ Dir( β ) end for for all documents m ∈ [1, M ] do − → → sample mixture proportion ϑ m ∼ Dir(− α) sample document length Nm ∼ P oiss(ξ) for all words n ∈ [1, Nm ] do − → sample topic index zm,n ∼ M ult( ϑ m ) → sample term for word wm,n ∼ M ult(− ϕ zm,n ) end for end for • M : the total number of documents • K: the number of (hidden/latent) topics • V : number of terms t in vocabulary − → → •− α , β : Dirichlet parameters − → • ϑ m : topic distribution for document m − → • Θ = { ϑ m }M m=1 : a M × K matrix − → • ϕ k : word distribution for topic k → • Φ = {− ϕ k }K k=1 : a K × V matrix • Nm : the length of document m • zm,n : topic index of nth word in document m • wm,n : a particular word for word placeholder [m, n]

all the other words. Technically, topic assignment of a word t is sampled from the following multinomial distribution. (t)

(k) nk,¬i + βt nm,¬i + αk → → p(zi = k|− z ¬i ,− w) = PV PK (v) (j) [ v=1 nk +βv ]−1 [ j=1 nm +αj ]−1 (1) (t) where nk,¬i is #times word t is assigned to topic k except PV (v) the current assignment; v=1 nk − 1 is the total #words (k) assigned to topic k except the current assignment; nm,¬i is the #words in document m assigned to topic k except the PK (j) current assignment; and j=1 nm − 1 is the total #words in document m except the current word t. After doing Gibbs Sampling, two matrices Φ and Θ are computed as follows. (t)

n + βt ϕk,t = PV k (v) v=1 nk + βv

(2)

(k)

nm + αk ϑm,k = PK (j) j=1 nm + αj

(3)

4.2. News Collection as Universal Dataset 4.2.1 Data preparation With the purpose of using a large scale dataset for Vietnamese contextual advertising, we chose VnExpress1 as the Universal Dataset. VnExpress is one of the highest ranking e-newspaper in Vietnam, thus contains a large number of articles in many topics in daily life. For this reason, it 1 VNExpress

e-newspaper: http://VnExpress.net/

is a suitable dataset for advertising area. The dataset includes different topics, such as Society, International news, Lifestyle, Culture, Sports, Science, etc. We crawled approximately 40,000 pages using Nutch2 . The data was preprocessed using JVnTextPro3 before being analyzed. This task includes: HTML removal, sentence segmentation, trivial or stop words removal, and word segmentation. As a Vietnamese word can contain more than one syllable, words are not always separated by white-spaces. Word segmentation is an important step to improve the overall accuracy. The statistics of the VnExpress dataset is given in Table 2. Table 2. VnExpress as Universal Dataset After removing html, doing sentence and word segmentation: size ≈ 219M, |docs| = 40, 328 After filtering and removing non-topic oriented words: size ≈ 53M, |docs| = 40, 268 |words| = 5, 512, 251; |vocabulary| = 128, 768

4.2.2 Topic Analysis of VnExpress Collection After preprocessing, the dataset was analyzed using GibbsLDA++4 . We carried out topic analysis of three models: 60, 120 and 200-topic, respectively. Some samples of hidden topics of 200-topic model are illustrated in Figure 3. The full lists of topics are available online5 .

the #words in document m assigned to topic k except the PK (j) current assignment; and j=1 nm − 1 is the total #words in document m except the current word t. After performing topic sampling, the topic distribu−→ tion of a new document − w→ m is ϑm = {ϑm,1 , . . . , ϑm,k , . . . , ϑm,K } where each ϑm,k is computed as follows: (k)

nm + αk ϑm,k = PK (j) j=1 nm + αj

Topics that have high probability ϑm,k will be added to the corresponding Web page/ad m. Each topic integrated into a Web page/ad will be treated as an external term and its frequency is determined by its probability value. Technically, the number of times a topic k is added to a Web page/ad m is decided by two parameters cut-off and scale: (

Frequencym,k =

Given an estimated LDA model as described in the previous section, we can now do topic inference for web pages and ads by a similar sampling procedure. In particular, we have a set of Web pages and ads W . Topic inference process will discover the probability distribution of topics over → − each document in W . Let − w and → z be the vectors of all words and their topic assignment of the whole new dataset → W . The topic assignment for a particular word t in − w depends on the current topic assignment for all the other words − → in → w and the topic assignment of all words in − w as follows: → − → − − → − → p(z i = k| z ¬i , w ; z ¬i , w ) = (t)

(t)

(k)

nm,¬i + αk nk,¬i + nk,¬i + βt PV PK (v) (v) (j) [ v=1 nk +nk βv ]−1 [ j=1 nm +αj ]−1

(4)

(t)

where nk,¬i is the #times word t is assigned to topic k PV − → (v) within W except the current assignment; v=1 nk − 1 − → is the total #words assigned to topic k in W that are as(k) signed to topic k except the current assignment; nm,¬i is 2 Nutch:

http://lucene.apache.org/nutch/ http://jvnsegmenter.sourceforge.net/ 4 GibbsLDA++: http://gibbslda.sourceforge.net 5 The full lists of hidden topics available at: gibbslda.sourceforge.net/ vnexpress-060topics.txt; vnexpress-120topics.txt; vnexpress-200topics.txt 3 JVnTextpro:

round (scale × ϑm,k ) , if ϑm,k ≥ cut-off 0, if ϑm,k < cut-off

where cut-off is the topic probability threshold. scale is a parameter that determines the topic frequency added. Title: Triệu trái tim (Million hearts) http://trieutraitim.info

Description: Trang_web (web page) nghe_nhạc (for music) giải_trí (entertainment) tổng_hợp_online (online) nghe (listen to) ca_khúc (song) nhạc (music) phim (film) tổng_hợp (collection) việt_nam (vietnamese) nước_ngoài (foreign) Keywords: nhạc_mp3 (mp3 music), nghe_nhạc (listen music) trực_tuyến (online), âm_nhạc (music), nhạc_trẻ (music for the youth), nhạc việt_nam (Vietnamese music)

Ad message before being enriched

5. Matching and Ranking with Hidden Topics 5.1. Topic Inference for Ads & Target Pages

(5)

a) Topic inference

b) Topics integrated into the ad Topic 57

Topic 104 ca_sĩ (singer) hát (sing) nhạc (melody) âm_nhạc (music) khán_giả (audience) nhạc_sĩ (composer) album (album) ca_khúc (song) chương_trình (program) sáng_tác (compose) sân_khấu (stage) biểu_diễn (perform) phong_cách (style) rock (rock) thể_hiện (show) giọng_hát (singing voice) ca_hát (singing) nghệ_sĩ (musician) ca_nhạc (harmony) diễn (act) viết (compose) fan (fan) tiếng (sound) thành_công (success) liveshow (live show) phát_hành (issue) chất (quality) thị_trường (market)

Triệu trái tim http://trieutraitim.info

Trang_web nghe_nhạc giải_trí Topic:104 tổng_hợp_online Topic:57 nghe những ca_khúc nhạc phim Topic:104 nhạc tổng_hợp việt_nam nước_ngoài nhạc_mp3 nghe_nhạc trực_tuyến Topic:104 âm_nhạc nhạc_trẻ nhạc việt_nam ca_nhạc Topic:104

Ad message a!er being enriched

Topics having high probabilies

Figure 4. An exp. of topic integration into ad An example of topic integration into ads is illustrated in Figure 4. The ad is about an entertainment Web site with a lot of music albums. After doing topic inference for this ad, hidden topics with high probabilities are added to its content in order to make it enriched and more topic-focused.

5.2. Matching and Ranking After being enriched with hidden topics, Web pages and ads will be matched based on their cosine similarity. For each page, ads will be sorted in order of its similarity to the page. The ultimate ranking function will also take into account the keyword bid information. But this is beyond the scope of this paper.

Topic 3

Topic 15

Topic 44

Topic 48

Topic 56

Topic 172

bác_sĩ (doctor) bệnh_viện (hospital) thuốc (medicine) bệnh (disease) phẫu_thuật (surgery) điều_trị (treatment) bệnh_nhân (patient) y_tế (medical) ung_thư (cancer) tình_trạng (condition) cơ_thể (body) sức_khoẻ (health) đau (hurt) gây (cause) khám (examine) kết_quả (result) căn_bệnh (illness) nặng (serious) cho_biết (inform) máu (blood) xét_nghiệm (test) chữa (cure) chứng (trouble)

thời_trang (fashion) người_mẫu (model) mặc (wear) trang_phục (clothes) thiết_kế (design) đẹp (beautiful) váy (dress) sưu_tập (collection) mang (wear) phong _cách (style) quần_áo (costume) nổi_tiếng (famous) quần (trousers) trình_diễn (perform) thích (like) quyến_rũ (charming) sang_trọng (luxurious) vẻ_đẹp (beauty) gái (girl) gương_mặt (figure) siêu (super) áo_dài (aodai) giày (shoes)

thiết_bị (equipment) sản_phẩm (product) máy (machine) màn_hình (screen) công_nghệ (technology) điện_thoại (telephone) hãng (company) sử_dụng (use) thị_trường (market) usd (USD) pin (battery) cho_phép (allow) samsung (Samsung) di_động (mobile) sony (Sony) nhạc (music) máy_tính (computer) hỗ_trợ (support) điện_tử (electronic) tính_năng (feature) kết_nối (connect) thiết_kế (design) chức_năng (function)

chứng_khoán (stock) công_ty (company) đầu_tư (investment) ngân_hàng (bank) cổ_phần (joint-stock) thị_trường (market) giao_dịch (transaction) đồng (VND) mua (buy) phát_hành (publish) niêm_yết (post) bán (sell) tài_chính (finance) đấu_giá (auction) trung_tâm (center) thông _tin (information) doanh_nghiệp (business) cổ_đông (shareholder) nhà _đầu_tư (investor) nhà _nước (government) tổ_chức (organization) triệu (million) quỹ (budget)

bánh (cake) mcdonald (McDonald) thịt (meat) pizza (pizza) ba_tê (pate) bánh_mì (bread) bánh_ngọt (pie) cửa_hàng (shop) xúc_xích (hot dog) kem (ice-cream) khai_trương (open) nguội (cold) hamburger (hamburger) thịt (meat) nhà _hàng (restaurant) đồ_ăn (food) sandwich (sandwich) khẩu_vị (taste) tiệm_bánh (bakery) bảo_đảm (ensure) nướng (grill) bí_quyết (secret) ngon (delicious)

thẻ (card) khoá (lock) rút (withdraw) chủ (owner) chìa (key) thẻ_tín_dụng (credit card) atm (ATM) tín_dụng (credit) thanh_toán (pay) visa (visa) tối_thiểu (minimum) mastercard phát_hành (release) trả_nợ (pay debt) sẵn_sàng (ready) mật_mã (password) thường_niên (annual) cảnh_giác (alert) chủ_thẻ (card owner) theo_dõi (follow) nhà _băng (bank) tội_phạm (criminal) trộm (steal)

Figure 3. Sample hidden topics and their most likely words from 200-topic model We verified the contribution of topics in many cases that normal keyword-based matching strategy cannot find appropriate ads for the target pages. Since normal matching is based on only the lexical feature of Web pages and ads, it is sometimes deviated by unimportant words that are not practical in matching. An example of such case is illustrated in Figure 5 (top 3 ads proposed by lexical matching method and matching with hidden topics method). The word “triệu” (million) is repeated many times in the target page, hence given a high weight in lexical matching. The system then misleads in proposing relevant ads for this target page. It puts ads having the same high-weighted word “triệu” in the top ranked list (c). However, those ads are totally irrelevant to the target page as the word “triệu” can have other meanings in Vietnamese. The words “chung_cư” (apartment) and “giá” (price) shared by top ads proposed by our method (Ad 21, 22, 23) and the target page on the other hands are important words although they do not have as high weights as the unimportant word “triệu” (f). However, by analyzing topics for them, we can find out their latent semantic relations and thus realize their relevance since they share the same topic 155 (g) and important words “chung_cư” (apartment) and “giá” (price). Topics analyzed for the target page and each ad are integrated to their contents as in Figure 5, b & e.

6. Evaluation 6.1. Experimental Data • For Web pages, we chose 100 pages randomly from a set of 27,763 pages crawled from VnExpress e-newspapers (exclusive from the Universal Dataset). Those pages were chosen randomly from different topics: Food, Shopping, Cosmetics, Mom & children, Estate, Stock, Jobs, Law, etc. These topics are primarily classified on the e-newspaper. And note that the information of these classified topics is not used in our experiments, just for reference here only.

• For advertisements, we collected 3,982 ads from Zing6 . Each ad is composed of four parts: title, Website’s URL, its description and keywords. All Web pages and ads were pre-processed before being matched. Finally, we had 100 pages and 2,706 unique ads for testing. They are available online7 for download.

6.2. Parameters & Evaluation Metrics Table 3. Description of 8 experiments Methods Without Hidden Topics

With Hidden Topics

AD AD_KW HT60_10 HT60_20 HT120_10 HT120_20 HT200_10 HT200_20

Description Use only title, description of ads Use title, description and keywords of ads number of topics = 60, scale = 10 number of topics = 60, scale = 20 number of topics = 120, scale = 10 number of topics = 120, scale = 20 number of topics = 200, scale = 10 number of topics = 200, scale = 20

To evaluate the contribution of hidden topics, we carried out six different experiments, which are called HT strategies. After doing topic inference for all Web pages and ads, we expanded their vocabularies with their most likely hidden topics. In the experiments, we used the value cutoff = 0.05 and tried two different scale: 10 and 20. The six matching experiments using hidden topics are called HT x_y, where x stands for the number of hidden topics of the used estimated model and y is the scale. We therefore performed six experiments: HT60_10, HT60_20, HT120_10, HT120_20, HT200_10, & HT200_20 (Table 3). To evaluate the performance of the matching method using retrieval information (term frequencies) only and the matching method using hidden topics, we prepared the test data following the methodology used in [11, 8]. First, we started by matching each Web page to all the ads and ranking them to their similarities. 8 methods proposed 8 different rank lists of ads to a target page. Since the 6 Vietnamese 7 Ad

Zing directory: http://directory.zing.vn/directory data: http://gibbslda.sourceforge.net/ContextualAd-TestData.zip

Target Page

a) Meanings of the target page and the ad messages

h!p://www.vnexpress.net/vietnam/kinhdoanh/bat-dong-san/2001/12/3b9b7a51/

Target page (Giá bán chung cư tái định cư…): an announcement of the selling prices for apartments, building, flat, etc.

Giá bán chung_cư tái_định_cư tuyến Lê Thánh Tôn nối dài chung_cư ngô_tất_tố lô a và lô b chung_cư Topic:155

Ad11 (Truyện tranh Tsubasa): an advertisement for Yoichi Takahashi and his very famous manga collection Captain Tsubasa Ad12 (Triệu trái tim): an entertainment web site with a lot of music albums, movies, and TV channels, etc. Ad13 (Ca sĩ Triệu Hoàng): personal homepage of singer Trieu Hoang

phạm_viết_chánh lô a và lô a1 đồng_giá tầng_trệt 3,86 triệu đồng m2 Topic:155 tầng_lửng lầu 1 đồng_giá triệu đồng lầu triệu đồng lầu 3 1,8 triệu lầu triệu lầu Topic:157 triệu đồng riêng lô c chung_cư Topic:157 ngô_tất_tố đơn_vị tính Topic:114 triệu đồng / m2

Ad21 (Công ty liên doanh Phú Mỹ Hưng): web site of Phu My Hung Corporation, one of the largest real estate companies in Vietnam Ad22 (Công ty Xây dựng và Thương mại đất Phương Nam): web site of Dat Phuong Nam Real Estate, Construction, and Trade Corporation Ad23 (Trung tâm giao dịch Bât động Sản Phúc Đức): web site of Phuc Duc Real Estate Investment Corporation

Giá bán chung cư... (Target Page)

10

Triệu trái tim (Ad 12)

20

triệu

Ca sĩ Triệu Hoàng (Ad 13)

8 6 4

Contribution of topics

12

Truyện tranh Tsubasa (Ad 11)

0

Shared topics

c)

d)

14

Giá bán chung cư... (Target Page)

12

Công ty liên doanh Phú Mỹ Hưng (Ad 21)

10

Công ty XD và TM Đất Phương Nam (Ad 22)

Công ty liên doanh Phú Mỹ Hưng -

Phương Nam - (Ad 22) www.datphuongnam.com.vn XâyTrung dựng,tâm phátgiao triểndịch nhà Bất phố,Động chungSản cư,Phúc cănĐức hộ. - (Ad 23) www.phucduc.com nhận ký_gởi rao_bán nhà Topic:155 villa chung_cư căn_hộ cao cấp đất nền dự_án đất dân_cư hiện hữu đất ao vườn ruộng

5

Shared words

b)

www.phumyhung.com.vn Topic:155 phát triển nhà, Công tyKinh Xâydoanh, dựng và Thương mại Đất

10

2 0

(Ad 21)

15

8

chung_cư

Trung tâm GD BĐS Phúc Đức (Ad 23)

6

giá

4

20

Topic 155 Contribution of topics

triệu trái_tim Topic:104 trang web nghe nhạc giải_trí tổng hợp online, Ca sĩ Triệu Hoàng – (Ad 13) nghe những ca khúc nhạc phim, nhạc tổng hợp, nhạc www.trieuhoang.com diễn_đàn ca_sĩ triệu hoàng Topic:104

Occurrence of words

14

Truyện tranh Tsubasa – (Ad 11) www.truyentranh.com/truyen_scan/subasa Subasa của tác_giả Topic:76 Yoichi Takahashi shounen đầu tiên được Triệu trái timlà–truyện (Ad 12) giới thiệu ở Việt Nam vào năm 1994, chuyện http://trieutraitim.info

Occurrence of words

Use both keywords & hidden topics

Match & Rank with keywords only

giá_như_sau

15

10

5

2 0

0

Shared words

f)

e)

Shared topics

g)

h) Topic 155 (most relevant to real estate & civil engineering) đất (land) xây_dựng (construction) dự_án (project) khu (area) thành_phố (city) quận (district) sử_dụng (usage) trung_tâm (central) đầu_tư (investment) công_trình (construction project) đô_thị (urban) khu_vực (local area) diện_tích (area) bất_động_sản (real estate) quy_hoạch (planning) người_dân (local people) đồng (rice field) xây (build) tầng (floor) tỉnh (province) nhà_ở (housing) căn_hộ (apartment) chung_cư (apartment building) phường (district) thuê (rent) dân (people) huyện (district) phát_triển (development) cho_thuê (for rent) rộng (wide) hạ_tầng (infrastructure) công_nghiệp (industry) chủ_đầu_tư (investor) môi_trường (environment) nằm (lie) triệu (million) tài_nguyên (resource) cho_biết (announce) tuyến (route) dân_cư (residents) mặt_bằng (area) dự_kiến (intend) tòa_nhà (building) kế_hoạch (plan) triển_khai (implement) cấp (issue) địa_bàn (local area) bán (sell) hoàn_thành (accomplish) kiến_trúc (architecture) địa_phương (local government) định_cư (settle) thương_mại (trade) kinh_doanh (business) thu_hồi (withdraw)

Figure 5. An example of matching and ranking without and with hidden topic inference number of ads is large, these lists can be different from this method to another method with little or no overlap. To determine the precision of each method and compare them, we selected top four ranked ads of each method and put them into a pool for each target page. Consequently, each pool will have no more than 32 ads. We then selected from these pools most relevant ads and exclude irrelevant ones. On average, each Web page will be matched with 6.51 ads eventually. The total number of Web pages is 100. To calculate the precision of each method, we used 11-point average score [9], a metric often used for ranking evaluation in IR.

6.3. Experimental Result & Analysis We used the method AD_KW as a baseline for our experiments using hidden topics. We examined the contribution of hidden topics using different estimated models: the model of 60, 120 and 200 topics. As illustrated in Figure 6 and Table 4, using hidden topics significantly improves the

performance of the whole framework. Figure 6 shows seven precision-recall curves of seven experiments in which the most inner line is the baseline and all the others are with hidden topics. From these curves, we can see the extent to which hidden topics can improve matching and ranking accuracy, and how the parameter values (i.e., number of topics, scale value) affect the performance. And from Table 4, we can see hidden topics help increase the precision on average from 64% to 72% and reduces almost 23% error (HT200_20). For the overall methods, we also calculated the number of corrected ads found in the first, second and third positions of the rank lists proposed by each strategy (#1, #2, #3 in Table 4). Because in contextual advertising, normally, we only consider some first ranked ads, we want to examine the precision of these top slots. It also reflects that the precision of our hidden-topic methods is higher than that of the baseline matching method. Moreover, the precision at position 1 (#1)

0.95

0.85

P rec is ion

0.75

0.65

AD_K W HT 60_10

0.55

HT 60_20 HT 120_10 HT 120_20

0.45

HT 200_10 HT 200_20

0.35 0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

R ec all

Figure 6. Precision-recall curves of matching & ranking without and with hidden topics is generally higher than that of position 2 and 3 (#2, #3). If the system is ranking the relevant ads near the top of the rank list, it is possible that the system can suggest most appropriate ads for the corresponding page. It therefore shows the effectiveness of the ranking system. Table 4. 11-points average precision Methods #1 AD AD_KW HT60_10 HT60_20 HT120_10 HT120_20 HT200_10 HT200_20

70 78 79 86 82 89 79 88

Correct ads found #2 #3 Totals 56 69 76 75 74 79 77 78

52 64 70 67 74 69 77 79

178 211 225 228 230 237 233 245

11-point avg. precision 49.86% 64.32% 68.72% 69.02% 70.76% 72.47% 70.26% 72.50%

Finally, we also quantified the effect of the number of topics and its added amount to each Web page and ad by testing with different topic models and adjusting the scale value. As indicated in Table 4, the performance of 120 and 200-topic models yields a better result than 60-topic model. However, there is no considerable change between 120-topic and 200-topic models, also in the quantities of added topics to each page and ad. It can therefore conclude that the number of topics should be large enough to discriminate the difference of terms to better analyze topics for Web pages and ads. And when the number of topics is large enough, the performance of the overall system become more stable.

7. Conclusions We have proposed a framework to rank most relevant ads for a Web page by taking advantage of hidden topics dis-

covered from a large data collection. This framework provides a mechanism to analyze topics for Web pages and ads, then integrate this information into their vocabularies. This helps overcome the problem of mismatch by capturing the semantic information and reducing the sparseness in the vocabularies of both Web pages and ads. The framework has shown its efficiency through a variety of experiments against the basic method using lexical information only. In practice, the results record an error reduction of 22.9 % in the method using 200-topic model over the normal matching strategy without hidden topics. Furthermore, it indicates that this high quality contextual advertising framework is easy to implement and practical in reality. We only need to collect a large-scale external data, which is available on the Web. Finally, the framework is also flexible and general enough to be applied in a multilingual environment. Other areas of future work include finding a ranking function to better describe the contribution of other features for finding most relevant ads beside their shared topics, such as keyword bid information and click-through rate.

Acknowledgements This work is partly supported by the research grant No.P06366 from Japan Society for the Promotion of Science (JSPS); by the Project QC.07.06 “Vietnamese Named Entity Recognition and Tracking on the Web”, Vietnam National University, Hanoi; and Vietnam National Project 02.2006.HĐ - ĐTCT-KC.01/06-10 “Web Content Classification, Clustering, and Filtering”.

References [1] D. Blei, A. Ng, and M. Jordan. Latent dirichlet allocation. Journal of Machine Learning Research, 3:993–1022, 2003. [2] A. Broder, M. Fontoura, V. Josifovski, and L. Riedel. A semantic approach to contextual ad. Proc. ACM SIGIR, 2007. [3] P. Chatterjee, D. L. Hoffman, and T. P. Novak. Modeling the clickstream: Implications for web-based advertising efforts. Marketing Science, 22(4):520–541, 2003. [4] R. Wang, P. Zhang, and M. Eredita. Understanding consumers attitude toward advertising. Proc. AMCIS, 2002. [5] T. Griffiths and M. Steyvers. Finding scientific topics. Proc. National Academy of Science, 101:5228–5235, 2004. [6] G. Heinrich. Parameter estimation for text analysis. TR. [7] T. Hofmann. Probabilistic LSA. Proc. UAI, 1999. [8] A. Lacerda, M. Cristo, M. Gonc¸alves, W. Fan, N. Ziviani, and B. Neto. Learning to advertise. Proc. ACM SIGIR, 2006. [9] C. Manning, P. Raghavan, and H. Schutze. Introduction to information retrieval. Cambridge University Press, 2008. [10] X.-H. Phan, L.-M. Nguyen, and S. Horiguchi. Learning to classify short and sparse text & web with hidden topics from large-scale data collections. Proc. WWW, 2008. [11] B. Neto, M. Cristo, P. Golgher, and E. de Moura. Impedance coupling in content-targeted advertising. Proc. SIGIR, 2005. [12] W. Yih, J. Goodman, and V. Carvalho. Finding advertising keywords on web pages. Proc. WWW, 2006.

Matching and Ranking with Hidden Topics towards ...

and an ad can still be a good match even when they share no common words .... contextual page-ad matching and ranking with hidden top- ics discovered ..... samsung (Samsung) di_đng (mobile) sony (Sony) nhc (music) máy_tính (computer).

491KB Sizes 3 Downloads 209 Views

Recommend Documents

Web Search Clustering and Labeling with Hidden Topics - CiteSeerX
Author's address: C.-T. Nguyen, Graduate School of Information Sciences, ...... stop); (2) the period can denote an abbreviation; (3) the period can be used.

Web Search Clustering and Labeling with Hidden Topics
There are three key requirements for such post-retrieval clustering systems: (1) .... —Providing informative and meaningful labels: Traditional labeling methods assume .... (2) use Latent Semantic Indexing (LSI) [Deerwester et al. 1990] to ..... Cu

Web Search Clustering and Labeling with Hidden Topics - CiteSeerX
relevant to the domain of application. Moreover ..... the support of hidden topics. If λ = 1, we ..... Táo (Apple, Constipation, Kitchen God), Chuô. t (Mouse), Ciju'a s.

Pricing and Matching with Frictions
The University of Chicago Press is collaborating with JSTOR to digitize, preserve and extend access to The ... Queen's University and Indiana University.

Matching with Myopic and Farsighted Players - coalitiontheory.net
Apr 19, 2017 - of the standard definition of the vNM stable set is that it does not take ..... It can be verified that the myopic#farsighted improving paths that were.

Bilateral Matching and Bargaining with Private Information
two$sided private information in a dynamic matching market where sellers use auctions, and ..... of degree one), and satisfies lim+$$. ; (2,?) φ lim-$$ ... matching technology is assumed to be constant returns to scale, it is easy to see that J(") .

Matching with Myopic and Farsighted Players - coalitiontheory.net
Apr 19, 2017 - heterogeneous with respect to their degree of farsightedness and confirm ... may result, with the top school enrolling up to 45 percent more high ...

Towards High-performance Pattern Matching on ... - Semantic Scholar
1Department of Automation, Tsinghua University, Beijing, 100084, China. ... of-art 16-MIPS-core network processing platform and evaluated with real-life data ...

Towards High-performance Pattern Matching on ... - Semantic Scholar
such as traffic classification, application identification and intrusion prevention. In this paper, we ..... OCTEON Software Developer Kit (Cavium SDK version 1.5):.

Towards Efficient Matching of Semantic Web Service Capabilities
facilitate Web services discovery and selection in the large network. In these .... the registry (in the aim of selecting the service that best fits the request) is equal.

Stable Matching With Incomplete Information
Lastly, we define a notion of price-sustainable allocations and show that the ... KEYWORDS: Stable matching, incomplete information, incomplete information ... Our first order of business is to formulate an appropriate modification of ...... whether

Matching with Contracts
electricity supply contracts.1 Package bidding, in which .... Milgrom model of package bidding, we specify ... Our analysis of matching models emphasizes.

Ranking with decision tree
This is an online mistake-driven procedure initialized with ... Decision trees can, to some degree, overcome these shortcomings of perceptron-based ..... Research Program of Chinese Academy of Sciences (06S3011S01), National Key Technology R&D Pro- .

Matching Problems with Expertise in Firms and Markets
Sep 3, 2009 - merger or notlwhat is the difference between good advice and advice that is ... undertake a merger. ..... theory of vertical and lateral integration.

Stability and Strategy-Proofness for Matching with ...
A Problem in the Japanese Medical Match and Its Solution. By Yuichiro ... 2Information about the matching program written ... which has a better incentive property. A ... h'eHr qh'. ·qh whenever qr < Ph'∈Hr qh' , and ¯qh = qh otherwise.

Bilateral Matching and Bargaining with Private ...
1455 de Maisonneuve Blvd. West, Montreal, Quebec H3G 1M8, Canada .School of International Business Administration,. Shanghai ... that when the discount rate is sufficiently small, the equilibrium is unique and has the property that every ...

Bilateral Matching and Bargaining with Private ...
two.sided private information in a dynamic matching market where sellers use auctions, and have ... used in some of the recent labor search literature, e.g. Kennan (2007). ..... Consider the situation where a type T buyer is in a proposing stage, giv

Substitutes and stability for matching with contracts - Science Direct
Jan 22, 2010 - www.elsevier.com/locate/jet. Substitutes and stability for matching with contracts. ✩. John William Hatfielda, Fuhito Kojimab,∗ a Graduate School of Business, Stanford University, Stanford, CA 94305, United States b Department of E

Stability and Strategy-Proofness for Matching with ...
Graduate Medical Education regulates the total number ... Cambridge,. MA. 02138, yka- [email protected]. Kojima: Department of. Economics, Stanford University, Stanford, CA 94305, ... 2Information about the matching program written.

Substitutes and Stability for Matching with Contracts
Sep 17, 2008 - Annual Conference at Waikiki, GAMES 2008 at Northwestern and ... the doctor-optimal stable mechanism is strategy-proof for doctors (Dubins ..... way. A matching problem with couples can be seen as a special instance of a ..... matching

Matching with Couples: Stability and Incentives in ...
to the 1983 match) it did not allow couples to report preferences over pairs of positions, one for each member of ... positions, because there does not necessarily exist a stable matching in markets with couples (Roth,. 1984). ...... Helppie, Brooke

gender discrimination estimation in a search model with matching and ...
discrimination and to show that, using the model and standard distributional assumptions, it .... action policy implemented as a quota system has no impact under the ...... on demographic characteristics, human capital, and life cycle issues. ..... B

Matching and Price Competition with an Outside Market
May 15, 2014 - Bulow and Levin (2006) develop a model of matching with salaries in which equilibrium salaries, relative to their competitive levels, are compressed and, on average, lower. This paper expands their model by allowing for a competitive o