Measuring Ad Effectiveness Using Geo Experiments Jon Vaver, Jim Koehler Google Inc.

Abstract

ods have been developed to quantify advertising’s incremental impact (see [1], [5], [3], [2]). Advertisers have a fundamental need to quan- Each method has its own set of advantages and tify the effectiveness of their advertising. For disadvantages. search ad spend, this information provides a Observational methods of measurement impose basis for formulating strategies related to bid- the least amount of disruption on an advertiser’s ding, budgeting, and campaign design. One ap- ongoing campaigns. In an observational study ad proach that Google has successfully employed to effectiveness is assessed by observing consumer measure advertising effectiveness is geo experi- behavior in the presence of the advertising over ments. In these experiments, non-overlapping a period of time. The analyses associated with geographic regions are randomly assigned to a these studies tend to be complex, and their recontrol or treatment condition, and each region sults may be viewed with more skepticism, berealizes its assigned condition through the use of cause there is no control group. That is, a statisgeo-targeted advertising. This paper describes tical model is used to infer the behavior of a comthe application of geo experiments and demon- parable set of consumers without ad exposure, strates that they are conceptually simple, have a as opposed to directly observing their behavior systematic and effective design process, and pro- via an unexposed control group. At Google, obvide results that are easy to interpret. servational methods have been used to measure the ad effectiveness of display advertising in the Google Content Network [1] and Google Search [2]. 1 Introduction The most rigorous method of measurement is a randomized experiment. One application of randomized experiments that is used to analyze search ad effectiveness is a traffic experiment. At Google, these are performed using the AdWords Campaign Experiments (ACE) tool [3]. In these experiments, each incoming search is assigned to a control or treatment condition and the subsequent user behavior associated with each condition is compared to determine the incremental impact of the advertising. These experiments are very effective at providing an understanding of consumer behavior at the query level. However, they do not account for changes in user behavior that occur further downstream from the search.

Every year, advertisers spend billions of dollars on online advertising to influence consumer behavior. One of the benefits of online advertising is access to a variety of metrics that quantify related consumer behavior, such as paid clicks, website visits, and various forms of conversions. However, these metrics do not indicate the incremental impact of the advertising. That is, they do not indicate how the consumer would have behaved in the absence of the advertising. In order to understand the effectiveness of advertising, it is necessary to measure the behavioral changes that are directly attributable to the ads. A variety of experimental and observation meth1

2

For example, conversion level behavior may involve multiple searches and multiple opportunities for ad exposures, and a traffic experiment does not follow individual users to track their initial control/treatment assignment or observe their longer-term behavior. An alternative approach is to vary the control/treatment condition at the cookie level. In a cookie experiment, each cookie belongs to the same control/treatment group across time. However, ad serving consistency is still a concern with cookie experiments because some users may have multiple cookies due to cookie churn and their use of multiple devices to perform online research. Cookie experiments have been used at Google to measure display ad effectiveness [5]. This paper describes one additional method for measuring ad effectiveness; the geo experiment. In these experiments, a region (e.g. country) is partitioned into a set of geographic areas, which we call “geos”. These geos are randomly assigned to either a treatment or control condition and geo-targeting is used to serve ads accordingly. A linear model is used to estimate the return on ad spend.

2

Geo Experiment Description

Online advertising can impact a variety of consumer behaviors. In this paper, we refer to the behavior of interest as the response metric. The response metric might be, for example, clicks (paid as well as organic), online or offline sales, website visits, newsletter sign-ups, or software downloads. The results of an experiment come in the form of return on ad spend (ROAS), which is the incremental impact that the ad spend had on the response metric. For example, the ROAS for sales indicates the incremental revenue generated per dollar of ad spend. This metric indicates the revenue that would not have been realized without the ad spend.

GEO EXPERIMENT DESCRIPTION

tiser, this region may be an entire country. There are two primary requirements for these geos. First, it must be possible to serve ads according to a geographically based control/treatment prescription with reasonable accuracy. Second, it must be possible to track the ad spend and the response metric at the geo level. Ad serving inconsistency is a concern due to finite ad serving accuracy, as well as the possibility that consumers will travel across geo boundaries. The location and size of the geos can be used to mitigate these issues. It is not generally feasible to use geos as small as, for example, postal codes. The generation of geos for geo experiments is beyond the scope of this paper. In the United States, one possible set of geos is the 210 DMAs (Designated Market Areas) defined by Nielson Media, which is broadly used as a geo-targeting unit by many advertising platforms. The next step is to randomly assign each geo to a control or treatment condition. Randomization is an important component of a successful experiment as it guards against potential hidden biases. That is, there could be fundamental, yet unknown, differences between the geos and how they respond to the treatment. Randomization ensures that these potential differences are equally distributed - statistically speaking across the treatment and control groups. It also may be helpful to constrain this random assignment in order to better balance the control and treatment geos across one or more characteristics or demographic variables. For example, we have found that grouping the geos by size prior to assignment can reduce the confidence interval of the ROAS measurement by 10%, or more.

Each experiment contains two distinct time periods: pretest and test (see Figure 1). During the pretest period there are no differences in campaign structure across geos (e.g. bidding strategy, keyword set, ad creatives, etc.). In this time period, all geos operate at the same baseline level and the incremental differences between the treatment and control geos in the ad spend and A geo experiment begins with the identification response metric are zero. of a set of geos, or geographic areas, that parti- During the test period the campaigns for the tion a region of interest. For a national adver2

Google Inc. Confidential and Proprietary

3

LINEAR MODEL

gate of the response metric during the pretest period for geo i, δi is the difference between the actual ad spend in geo i and the ad spend that would have occurred without the experiment, and i is the error term. This model is fit using weights wi = 1/yi,0 in order to control for heteroscedasticity caused by the differences in geo size. The first two parameters in the model, β0 and β1 , are used to account for seasonal differences in the response metric across the pretest and test periods. The parameter of primary interest is β2 , which is the return on ad spend (ROAS) of the response metric.

Figure 1: Diagram of a geo experiment. Ad spend is modified in one set of geos during the test period, while it remains unchanged in another. There may be some delay before the corresponding change in a response metric is fully The values of yi,1 and yi,0 (e.g. offline sales) are realized generated by the advertiser’s reporting system. The geo level ad spend is available through Adtreatment geos are modified. This modification Words. If there is no ad spend during the pretest generates a nonzero differential in the ad spend period then the ad spend differential, δi , required in the treatment geos relative to the control geos. by Equation 1 is simply the ad spend during the That is, the ad spend differs from what it would test period. However, if the ad spend is positive have been if the campaign had not been modiduring the pretest period and is either increased fied. This differential will be negative if the camor decreased, as depicted in Figure 1 , then the paign change causes the ad spend to decrease ad spend differential is found by fitting a second in the treatment geos (e.g. campaigns turned linear model: off), and positive if the change causes an increase si,1 = γ0 + γ1 si,0 + µi (2) in ad spend (e.g. bids increased or keywords added). This ad spend differential will gener- Here, s is the ad spend in geo i during the test i,1 ate a corresponding differential in the response period, s is the ad spend in geo i during the i,0 metric, perhaps with some time delay, ν. Offline pretest period, and µ is the error term. This i sales is an example of a response metric that is model is fit with weights w = 1/s using only i i,0 likely to have a positive value of ν. It takes time the control geos (C). for consumers to complete their research, make a decision, and then visit a store to make their This ad spend model characterizes the impact of purchase. The test period extends beyond the seasonality on ad spend from the pretest period end of the ad spend change by ν to fully capture to the test period, and it is used as a counterfactual 1 to calculate the ad spend differential. The these incremental sales. ad spend differential in the control and treatment geos (T ) is found using the following prescription: 3 Linear Model  si,1 − (γ0 + γ1 si,0 ) for i ∈ T δi = (3) 0 for i ∈ C After an experiment is executed, the results are analyzed using the following linear model: The zero ad spend differential in the control geos yi,1 = β0 + β1 yi,0 + β2 δi + i

(1)

reflects the fact that these geos continue to operate at the baseline level during the test period.

1 where yi,1 is the aggregate of the response metric The counterfactual is the ad spend that would have during the test period for geo i, yi,0 is the aggre- occurred in the absence of the treatment.

Google Inc. Confidential and Proprietary

3

5

4

Example Results

DESIGN

Return On Ad Spend for Clicks

One of Google’s advertisers ran an experiment to measure the effectiveness of their search advertising campaign. During this experiment, which lasted several weeks, the advertiser’s search ads were shown in half of the geos. Figure 2 shows the result of fitting the linear model in Equation 1 with successively longer sets of test period data to find the ROAS for clicks. At first, the confidence interval of this metric is large, but it decreases quickly as more test period data are accumulated. Each dollar of ad spend generates 1/3 of an incremental click or, equivalently, the CPIC is $3. In this case, the reported CPC in AdWords is $2.40, which underestimates CPIC by 20% 2 . So, the paid clicks do displace some organic clicks, but certainly not the bulk of them.

0.4



● ●

●●

●●●

●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●

0.3

●●●●

0.2

end of incremental ad spend

0.1

ROAS for Clicks

ROAS 95% conf. interval conf. interval width



0.0

One issue that is of primary concern to advertisers is the potential cannibalization of cost-free organic clicks by paid search clicks (i.e. users will click on a paid search link when they would have clicked on an organic search link). Although perhaps unlikely, it is also possible that the cooccurrence of a paid link and an organic link will make an organic click more likely. Cost per click (CPC) does not provide the advertiser with a complete picture of advertising impact because of competing effects such as these. A more useful metric is the cost per incremental click (CPIC), which can be measured with a geo experiment.

0.5

0.6



0

10

20

30

40

50

Time Since Test Start

Figure 2: Measurement of return on ad spend for clicks as a function of test period length. The uncertainty in this estimate decreases until the ad spend returns to normal levels in all of the geos. that, in this case, search advertising does not increase the number of clicks beyond the day in which the ad spend occurred.

As mentioned in Section 2, the impact of ad spend is not as time-limited for all response metrics. Figure 4 is analogous to Figure 3, except the response metric is offline sales. Even after the ad spend differential returns to normal, the impact of the ad spend continues to generate incremental sales for some period of time before To further illustrate the ability of paid search fading. advertising to generate incremental clicks, Figure 3 shows the cumulative incremental ad spend across the test period along with the cumulative 5 Design incremental clicks. The number of incremental clicks is zero at the beginning of the test period and increases steadily with time along with Design is a crucial aspect of running an effective the incremental ad spend. However, once the ad geo experiment. Before beginning a test, it is spend in the test geos returns to a pretest level, helpful to understand how characteristics such as the accumulation of incremental ad spend stops. experiment length, test fraction, and magnitude At the same time, the accumulation of incremen- of ad spend differential will impact the uncertal clicks stops as well. This behavior indicates tainty of the ROAS measurement. This understanding allows for the design of an effective and efficient experiment. Fortunately, it is possible 2 In [2] the authors define IAC (incremental ad clicks) as the fraction of paid clicks that are incremental. IAC = to make such assessments for the linear model in Equation 1. CPC / CPIC, so IAC = 80% in this example. 4

Google Inc. Confidential and Proprietary

5

DESIGN

250000

cumulative incr. spend cumulative incr. clicks

150000

var(β2 ) = 

1−

ρ2yδ

σ2  hP 

N i=1 wi (δi

¯2 − δ)

i

(4)

where σ is the residual variance, and i2 ¯ − y¯0 )(δi − δ) (5) = PN P ¯2 ¯0 )2 N i=1 wi (δi − δ) i=1 wi (yi,0 − y hP

ρ2yδ

50000

Cumulative Incremental Ad Spend & Clicks

Cumulative Incremental Ad Spend & Clicks

For anP experiment with N geos, let y¯0 = ¯ = (1/N ) PN δi . Lin(1/N ) N y and δ i,0 i=1 i=1 ear theory indicates that the variance of β2 from Equation 1 is

end of incremental ad spend

N i=1 wi (yi,0

0 1200000







ad spend









800000

● ● ●

400000





cumulative incr. spend cumulative incr. revenue





0

Cumulative Incremental Ad Spend & Revenue

(see Appendix) . Using a set of geo-level pretest data in the response variable, it is possible to 0 10 20 30 40 50 use this expression to estimate the width of the Time Since Test Start ROAS confidence interval for a specified design Figure 3: Cumulative incremental ad spend and scenario. clicks across the test period. The accumulation of incremental clicks stops as soon as the ad The first step in the process is to select a consecutive set of days from the pretest data to create spend returns to the pretest level in all geos. pseudo pretest and test periods. The lengths of the pseudo pretest and test periods should match the lengths of the corresponding periods in the hypothesized experiment. For example, an experiment with a 14 day pretest period and a 14 day test period should have pseudo pretest and Cumulative Incremental Ad Spend & Revenue test periods that are each 14 days long. The end of data from the pseudo pretest period are used to incremental estimate yi,0 and wi in Equation 4.



0



5

10

15

The next step is to randomly assign each geo to the treatment or control group. We have found that confidence interval estimates are lower by about 10% when this random assignment is constrained in the following manner. The geos are ranked according to yi,0 . Then, this ranked list of geos is partitioned into groups of size M , where the test fraction is 1/M . One geo from each group is randomly selected for assignment to the treatment group.

Time Since Test Start

Figure 4: Cumulative incremental sales across the length of the test period. Incremental sales continue to be generated even after the ad spend returns to pretest levels in all geos.

Google Inc. Confidential and Proprietary

It may be possible to directly estimate the value of δi at the geo level. For example, if the ad spend will be turned off in the treatment geos, then δi is just the average daily ad spend for treatment geo i times the number of days in the experiment. Otherwise, an aggregate ad spend differential ∆ can be hypothesized and the geo5

6

level ad spend differential can be estimated using for i ∈ T for i ∈ C

(6)

The last value to estimate in Equation 4 is σ . This estimate is generated by considering the reduced linear model; yi,1 = βˆ0 + βˆ1 yi,0 + ˆ

(7)

0.15

i yi,0 )

Analysis Results Prediction 0.10

∆(yi,0 / 0

Confidence Interval Prediction

0.05

δi =

P

Confidence Interval Half Width (95%)



CONCLUDING REMARKS

begin multiple use of pretest data

end of incremental ad spend

0.00 0.0

This model has the same form as Equation 1 except the ad spend differential term has been 0 10 20 30 40 dropped. Fitting this model using the pseudo Time Since Test Start pretest and test period data results in a residual variance of σˆ, which is used to approximate σ . Figure 5: ROAS confidence interval prediction To avoid any peculiarities associated with a par- across the length of the test period. The predicticular random assignment, Equation 4 is evalu- tion is quite good until the test period becomes ated for many random control/treatment assign- long enough that some of the pretest data must ments. In addition, different partitions of the be used multiple times to generate each estimate pretest data are used to create the pseudo pretest of var(β2 ). and test periods by circularly shifting the data in time by a randomly selected offset. The half length of the hypothesized pretest and test periwidth q estimate for the ROAS confidence interval ods becomes longer than the (deliberately) limis 2 var(β2 ), where var(β2 ) is the average vari- ited set of pretest data used to generate the esance of β2 across all of the random assignments. timates. The good match between these two This process can be repeated across a number of curves demonstrates that the absolute size of the different scenarios to evaluate and compare de- confidence interval can be predicted quite well, signs. Note that if a limited set of pretest data at least as long as the ad spend differential can is available, circular shifting of the data makes be accurately predicted. it possible to analyze scenarios with extended test periods. However, doing so requires data points to be used multiple times in generating 6 Concluding Remarks each estimate of var(β2 ), and the example below demonstrates that this reuse of the data leads to Measuring ad effectiveness is a challenging probestimates that are overly optimistic. lem. Currently, there is no single methodology Figure 5 shows the confidence interval predic- that works well in all situations. However, geo tion as a function of experiment length for the experiments are worthy of consideration in many click example from Section 2. The dashed line situations because they provide the rigor of a corresponds to the predicted confidence inter- randomized experiment, they are easy to underval half width and the solid line corresponds stand, they provide results that are easy to into results from the experiment. For this com- terpret, and they have a systematic and effective parison, the ad spend differential from the ex- design process. Geo experiments can be applied periment was used as input to the prediction. to measure a variety of user behavior and can The predictions are quite accurate beyond the be used with any advertising medium that alvery beginning of the test period. Additionally, lows for geo-targeted advertising, Furthermore, they maintain this accuracy until the combined these experiments do not require the tracking of 6

Google Inc. Confidential and Proprietary

7

APPENDIX

individual user behavior over time and therefore these translations, the relevant linear model beavoid privacy concerns that may be associated comes 0 0 yi,1 = β1 yi,0 + β2 δi0 + i (8) with alternative approaches. Or, Y = Xβ + 

Acknowledgments

(9)

where We thank those who reviewed this paper (with special thanks to Tony Fagan and Lizzy Van Alstine for their many helpful suggestions), others at Google who made this work possible, and the forward looking advertisers who shared their data with us.

 0  0  y1,1 y1,0 δ10  .   . .         .  Y =  . , X =  .   .   . .  0 0 0 yN,1 yN,0 δN 



References

β=

  β1 , β2

 1  .     =  .   .  N

[1] D. Chan, et al. “Evaluating Online Ad Campaigns in a Pipeline: Causal Models at With the model in this form, the varianceScale.” Proceedings of ACM SIGKDD 2010, covariance matrix of the weighted least squares pp. 7-15. estimated regression coefficients is: [2] D. Chan et al. “Incremental Clicks var(β) = σ2 (X T W X)−1 (10) Impact Of Search Advertising.” research.google.com/pubs/archive/37161.pdf, (see [4]), where W is a diagonal matrix containing the weights wi , 2011.   w1 0 . . 0 [3] Google Ads Team. “AdWords Campain Ex 0 w2 .  periments.” Sept. 1, 2011. Ad Innovations.   . .  (11) W= .  . http://www.google.com/ads/innovations/   . . 0 ace.html 0 . . . wN [4] M. H. Kutner, et al. Applied Linear Statistical Models. New York: McGraw-Hill/Irwin, Now, "P #−1 2005. P 0 2 0 δ0 wi yi,0 2 P i wi yi,0 1 i P (12) var(β) = σ 0 0 02 [5] T. Yildiz, et al. “Measuring and i wi yi,0 δi i wi δi Optimizing Display Advertising Impact Through Experiments” In preparation (re- and the last component of this matrix is the variance of β2 , search.google.com) P 0 2 σ2 i wi yi,0 var(β2 ) = P  P  P 2 . 02 − 0 δ0 0 2 7 Appendix w δ w y w y i i i,0 i i i i,0 i i i (13) To derive Equation 4, consider the centered ver- Using Equation 5, sions of the variables yi,1 , yi,0 , and δi from Equa! ! X X 0 = y 0 = y 0 = tion 1; yi,1 − y ¯ , y − y ¯ , and δ 2 2 0 i,1 1 i,1 i,0 i wi δi0 (1 − ρyδ )2 = wi yi,0 P0 ¯ δi − δ for i ∈ 1...N and y¯j = (1/N ) i yi,j . With i i Google Inc. Confidential and Proprietary

7

7

! X i

0 2 wi yi,0

!2

! X i

2 wi δi0

APPENDIX



X

0 wi yi,0 δi0

i

(14) which, after substituting into Equation 13, leads to Equation 4.

8

Google Inc. Confidential and Proprietary

Measuring Ad Effectiveness Using Geo ... - Research at Google

website visits, newsletter sign-ups, or software downloads. The results of an ..... search.google.com/pubs/archive/37161.pdf,. 2011. [3] Google Ads Team.

311KB Sizes 1 Downloads 408 Views

Recommend Documents

Estimating Ad Effectiveness using Geo ... - Research at Google
situations, the development of alternative analysis methodologies to provide advertisers ... Nielsen Company, 2017), were originally formed based on television viewing behavior of .... A geo experiment was executed in the U.S. using all 210 DMAs, whi

Measuring Ad Effectiveness Using Geo Experiments - Semantic Scholar
does not follow individual users to track their initial control/treatment assignment or observe their longer-term ... their use of multiple devices to perform online research. Cookie experiments have been used at. Google to ... Second, it must be pos

Measuring Interference Between Live ... - Research at Google
generic measurement techniques for analyzing interference in the face of these .... a prohibitively expensive way to find baseline performance, especially given ..... runs with the negative co-runner and call it the “negative frac- tion”. Call th

MEASURING NOISE CORRELATION FOR ... - Research at Google
the Fourier frequency domain. Results show improved performance for noise reduction in an easily pipelined system. Index Terms— Noise Measurement, Video ...

Mesa: Geo-Replicated, Near Real-Time ... - Research at Google
Jan 1, 2014 - [10] have proposed the MaSM (materialized sort-merge) algorithm, which can be used ... lems of data corruption that may result from software errors and hardware ...... other companies that need to be supported. In summary,.

Geo-level Bayesian Hierarchical Media Mix ... - Research at Google
shape effect, media spend should be standardized to the amount per capita. In some use cases, .... For simplicity of illustration, we will use geometric adstock x∗.

The Unreasonable Effectiveness of Data - Research at Google
Feb 5, 2010 - Contact Editor: Brian Brannon, [email protected] such as f = ma or e ... comes to natural language processing and related fields, we're ...

Best Practices for Maximizing Effectiveness - Research at Google
http://uxmanagement.wikispaces.com/UX. About the Authors. Jhilmil Jain is a senior UX research manager at Android with a team that spans the U.S. and the.

Measuring Advertising Quality on Television - Research at Google
Dec 3, 2009 - they reported on the five best-liked ads and the five most-recalled ads. ... audience behavior. By adjusting for features such as time of day, network, recent user .... TV network are included but not the specific campaign or ... chose

One Billion Word Benchmark for Measuring ... - Research at Google
amount of data involves a large amount of work, and provides a significant barrier to entry for new mod- eling techniques. By choosing one billion words as.

Measuring Cross-Device Online Audiences - Research at Google
May 10, 2016 - The method performs demographic corrections in the usual way device-by-device. ... of online ad campaigns by audience attributes for one device (or cookie) type. ..... 0|c)=1 − e−κ0c/P . We call this cookie-to-user function the Di

An Active Approach to Measuring Routing ... - Research at Google
studied by analyzing routing updates collected from the pub- lic RouteViews ..... the RRCs using the announcer [9] software developed by one of the authors.

How Surfers Watch: Measuring audience ... - Research at Google
Jul 25, 2010 - viewer response to video ad creatives more generally. 1. ... are the effect due to the creator of the underlying video programming (the ...

Measuring User Rated Language Quality ... - Research at Google
Items 1 - 9 - .360 .616 .257 .431 .811. Google AdWords *. 400 .670 .900 .368 .632 .249 .386 .809. Note. pv = item ..... Missing data: our view of the state of the art.

Indirect Content Privacy Surveys: Measuring ... - Research at Google
We present a design for indirect surveys and test the design's use as (1) a means to ... concerned about company/government access to health records. [39]. Almost as ... their purchase histories on web sites like Blippy [4], are will- ing to trade ..

Measuring Enforcement Windows with Symbolic ... - Research at Google
Static analysis tools, which help find bugs in software, can take advantage of isolation boundaries to scale to real-world programs. Can we find and leverage the ...

A Method for Measuring Online Audiences - Research at Google
We present a method for measuring the reach and frequency of online ad ... is that a cookie does not identify a person, but a combination of a user account, ..... Analysis of Complex Survey Samples Journal of Statistical Software, 9(1), 1-19,.

Content Fingerprinting Using Wavelets - Research at Google
Abstract. In this paper, we introduce Waveprint, a novel method for ..... The simplest way to combine evidence is a simple voting scheme that .... (from this point on, we shall call the system with these ..... Conference on Very Large Data Bases,.

SOUND SOURCE SEPARATION USING ... - Research at Google
distribution between 0 dB up to clean. Index Terms— Far-field Speech Recognition, Sound ... Recently, we observed that training using noisy data gen- erated using “room simulator” [7] improves speech ... it turns out that they may show degradat

Near Impressions for Observational Causal Ad ... - Research at Google
... the performance of their online advertising by either running ran- ..... sure the impact of their display and video ad ..... iment” Journal of Educational Psychology.

A Truthful Mechanism for Offline Ad Slot ... - Research at Google
Jan 18, 2008 - 1 Google, Inc. Email: {jonfeld,muthu,mpal}@google.com ... interpretation in this application: if an overall advertising campaign allocates a fixed portion of its budget ... truthful mechanism (see Example 1 in Section 2). However ...

Evaluating Online Ad Campaigns in a Pipeline - Research at Google
KDD'10, July 25–28, 2010, Washington, DC, USA. Copyright 2010 ACM ... may simultaneously run similar campaigns in other media such as ..... demiology and the social sciences. .... verse propensity weighting is best in the sense of smallest.

Ad Quality On TV: Predicting Television ... - Research at Google
Jun 28, 2009 - ABSTRACT. This paper explores the impact of television advertisements on audience retention using data collected from television set-top ...

Measuring the effectiveness of personal database ...
Our subjects clearly had differing ideas about the physical .... business life deception habit health help influence judgement learn listening love junk. manners .... display device employing several virtual screens for presentation of one or more.