Proceedings of International Joint Conference on Neural Networks, Montreal, Canada, July 31 - August 4, 2005

Estimation of Prediction Intervals for the Model Outputs Using Machine Learning D. L. Shrestha

D. P. Solomatine

UNESCO-IHE Institute for Water Education P.O. Box 3015, 2601 DA Delft, The Netherlands E-mail: [email protected]

UNESCO-IHE Institute for Water Education P.O. Box 3015, 2601 DA Delft, The Netherlands E-mail: [email protected]

Abstract− A new method for estimating prediction intervals for a model output using machine learning is presented. In it, first the prediction intervals for insample data using clustering techniques to identify the distinguishable regions in input space with similar distributions of model errors are constructed. Then regression model is built for in-sample data using computed prediction intervals as targets, and, finally, this model is applied to estimate the prediction intervals for out-of-sample data. The method was tested on artificial and real hydrologic data sets using various machine learning techniques. Preliminary results show that the method is superior to other methods estimating the prediction intervals. A new method for evaluating performance for estimating prediction intervals is proposed as well.

used in statistical [2; 11] and machine learning communities [3; 12; 13] for time series forecasting. In this approach uncertainty is estimated in terms of confidence intervals or prediction intervals. The third approach is to use simulation and re-sampling based techniques, generally referred to as ensemble or Monte Carlo methods (one of the versions of such approach used in hydrologic modeling is a Generalized Likelihood Uncertainty Estimator, GLUE [7]). The first and the third approaches mentioned require the prior distributions of the uncertain input parameters or data to be propagated through the model to the outputs. While the second approach requires certain assumptions about the data and the errors and obviously the relevancy and accuracy of such approach depends on the validity of these assumptions. In this paper, we propose a new approach for estimating uncertainty of the model outputs. In the situations when distributions of parameters or distributions of model errors are not known, we propose to use Machine Learning (ML) methods to quantify uncertainty of the model outputs in the form of Prediction Intervals (PIs). In our experiments, we used linear regression, Locally Weighted Regression (LWR), and M5 model trees (MT) as the Prediction Intervals Model (PIM) using the all or part of input vectors that are used to train (calibrate) the main forecasting models. Upper and lower PIs are computed separately on the basis of errors observed during training of the forecasting models. Fuzzy C-means clustering technique is used to determine zones of the input space having similar distributions of model errors. The PIs of each cluster are determined on the basis of empirical distributions of the errors associated with all instances belonging to this cluster. The PIs of each instance are then determined according to grade of their membership in each cluster. After training the PIMs, they are applied to estimate the PIs for the forecasts made by three machine learning techniques (linear regression, LWR, and MT) in artificial and real hydrologic data sets. The performance of the new method is compared to other methods.

I. INTRODUCTION In solving the problems of model-based forecasting of natural phenomena, for example, in flood management, decision makers require not only accurate forecasting of certain variables but also the uncertainty estimates associated with the forecasts. In weather prediction the forecasts are typically given together with the associated uncertainty, but in other areas, e.g., in water management the prevailing format of forecasting is deterministic, i.e. in the form of a point forecast. It does not take into account the various sources of uncertainties including model uncertainty, input uncertainty and parameter uncertainty. Incorporating prediction intervals into deterministic forecasts explicitly takes into account all sources of uncertainty of the model outputs and thus helps to enhance the reliability and credibility of the model outputs. Uncertainty of the model output can be estimated using several approaches. The first approach is to forecast the model outputs probabilistically and it is often used in hydrological modelling [9]. The second approach is to estimate uncertainty by analyzing the statistical properties of the model errors that occurred in reproducing the observed historical data. This approach has been widely

0-7803-9048-2/05/$20.00 ©2005 IEEE

2700

(often called the Prediction Mean Squared Error (PMSE)

II. PREDICTION INTERVALS An interval forecast is usually comprised of the upper and lower limits between which a future unknown value (e.g. point forecast) is expected to lie with a prescribed probability. These limits are called prediction limits or bounds, while the interval is called Prediction Intervals (PIs). The prescribed probability is called confidence level. PIs should be distinguished from Confidence Intervals (CIs). The CIs deal with the accuracy of our estimate of the true regression (statisticians formulate this differently – as estimation of the mean value of the outputs [2; 11]). In contrast, the PIs deal with the accuracy of our estimate with respect to the observed target value. It should be noted that the PIs are wider than the CIs [12]. In real world applications, the PIs are of more practical use than CIs. This is because prediction interval are concerned with the accuracy with which we can predict the observed targets values themselves, but not just the accuracy of our estimate of the true regression.

[1]). However, σ e2

is not known in practice and is

estimated from the data. An unbiased estimate of σ e2 with N − p degrees of freedom, denoted by se2 , is given by the formula se2 = SSE / ( N - 2) N

N

i =1

i =1

(3)

where SSE = ∑ ei2 = ∑ (ti − yi ) 2 is the Sum Squared Error, and N is number of data. Furthermore, eq. (2) assumes that the error distribution is normal with zero mean. If the error variance σ e2 is not constant, then eq. (3) should be modified to take into account this effect. Modified value of se2 at x = x0 denoted by s 2f , is given by the following formula [2; 11]:

The following sub-sections present briefly the methods for constructing PIs for the model outputs.

s 2f ( x0 ) = se2 (1 + 1 / N + ( x0 − x ) 2 /( N − 1) s x2 )

(4)

A. An Approach for Linear Regression Models

where s x2 is the sample standard deviation of x , and x is

To simplify the formulation, let us assume that the forecasting model is a univariate linear regression model, whose aim is to obtain forecasts of the dependent variable t given some values of the independent variable x. Let the forecast of t be y, which may not necessarily be the same due to various reasons including presence of noise in the data, limited number of data points, non-linear relationship between the dependent and independent variables, and the errors in estimating parameters. So there is a discrepancy between the target value and the forecasted value by the regression equation (model), which is simply the error of the forecast denoted by e=t− y

the sample mean of x . It can be seen that s 2f is always

(1)

where y = a + bx , and a, b are the parameters of the regression. Most of the methods to construct 100(1- α )% prediction limits for t used in practice are essentially of the following form: PLU = y + z α / 2 σ e

(2)

PLL = y − z α / 2 σ e

where PLU and PLL are the upper and lower prediction limits respectively, z α / 2 is the value of the standard normal variate with cumulative probability level of α / 2 , and σ e is the standard deviation of the errors. Since the prediction limits in eq. (2) are symmetric about y, it is assumed that the forecast is unbiased. Consequently, the forecast error variance σ e2 equals to the Mean Squared Error (MSE)

larger than se2 . In addition, s 2f is dependent on how far x0 is away from x . This methed will be referred to as Linear Regression Variance Estimator (LRVE). B. Resampling Approach For non-linear models, eq. (4) can not be applied to estimate σ e2 and the derivation of σ e2 for such non-linear models is not so easy. However, resampling based techniques can be used to estimate σ e2 and thus to compute the PIs; these are often referred to as error bars and are used, for example, in neural networks research [3; 12; 13]. Typically, these techniques are based on the premise that the error variance σ e2 can be decomposed into three terms; model bias, model variance, and target noise [10]. Model variance can be estimated by a committee (ensemble) of networks, while the target noise is estimated by training an auxiliary network on the residuals of the committee predictions. Most of such techniques neglect the contribution of model bias to the total error variance. These techniques are generally computationally time consuming, as the ensemble should include many networks to ensure a reliable estimate. The description of such techniques is beyond the scope of this paper. III. METHODOLOGY LRVE approach requires many assumptions to apply eq. (4) to construct PIs for linear regression. Furthermore, such

2701

theoretical methods are difficult or impossible to use, especially in case of non-linear or multivariate models containing many equations. For non-linear problems, there exist few resampling based methods, which computational cost is very high. This section describes new method for estimating uncertainty of model outputs in terms of the PIs. Due to the various sources of uncertainty as mentioned in the section I, it is not surprising that model outputs do not match the observed values well. The essence of the proposed method is that historical residuals (errors) between the model outputs and the corresponding observed data are the best available quantitative indicators of the discrepancy between the model and the modeled real world system or process. These residuals between the model outputs and observed values give the valuable information that can be used to assess the model uncertainty. These residuals are often function of the model input’s values and can be modeled. ∑ µi N

∑ µi

i =1 N

(1 − α / 2) ∑ µi

respect to the corresponding errors in ascending order. The following expression gives the lower PI for cluster j, ( PIC L ): PIC L ( j ) = e(i )

(5)

i

N

k =1

k =1

i : ∑ µ j (k ) < α / 2 ∑ µ j (k )

where i is the maximum value of it that satisfies the above inequality, e(i) is the error (residual) associated with the instance i (instances are sorted) , µ j (i ) is the membership grade of ith instance to cluster j. Similar type of expression can be obtained for the upper PI ( PICU ). This is illustrated in Fig. 1. Once the PIs are computed for each cluster, the PIs for each instance in input space can be computed. Computation of the PIs for each instance also depends upon the clustering techniques employed. For example, if K-means clustering is employed, then the PIs for each instance in the particular cluster are the same as that of the cluster. However, if fuzzy C- means clustering is employed, PIs are computed using the weighted mean of PIs of each cluster as:

i =1

c

PI L (i ) = ∑ µ j (i ) PIC L ( j )

(6)

j =1

N

α / 2 ∑ µi i =1

-2.65 -1.56

-1.45 -1.45

-0.23

Lower interval

0.58

1.12

1.13

1.85

2.01

2.69

e

Upper interval

Fig. 1. Computation of prediction intervals in case of using fuzzy C-means clustering

The method identifies n zones (regions) in input space reproducing different distributions of historical residuals. With each region can be associated a cluster of input vectors. It is assumed that a region in input space that is associated with any particular cluster has similar residuals or residuals with similar distributions. Having identified the clusters, the PIs for the particular cluster are computed from empirical distributions of the historical residuals that belong to the cluster under consideration. For instance, in order to construct 100(1- α )% PIs, α / 2 *100 and (1 − α / 2) *100 percentile values are taken from empirical distribution of residuals for lower and upper PIs respectively. Typical value for α is 0.05, so this corresponds to 95% interval. If input space is divided into crisp clusters, e.g., by K-means clustering method, and each instance belongs to exactly one cluster, this computation is straightforward. However, in case of fuzzy clustering, e.g., by fuzzy C-means method [6], where each instance belongs to more than one cluster and is associated with several membership grades, the computation of the above percentiles should take this into account. To calculate PIs, the instances should first be sorted with

where PI L (i ) is the lower PI for ith instance, c is number of clusters. Similar type of expression can be obtained for the upper PI ( PIU ). Once the PIs for each input instance is computed, it is possible to construct mapping function f PI that estimate underlying functional relationship between input x and target PIs as: PI = f PI (x)

(7)

where PI without subscripts denotes PI in general. Mapping function f PI may be of any form, from linear regression to non-linear functions such as neural networks. In other words, given a set of N data pairs {x(n), PI(n)}, n = 1,…, N, we can train a neural network or another regression model to estimate underlying function relating x to PI . It is to be noted that the target variable might be either intervals or limits. Mapping function f PI will be referred to as the Prediction Intervals Model (PIM). IV. VALIDATION It is often difficult to validate a PIM model models due to unavailability of the necessary data. For instance, we have PIs as targets in in-sample (training) data, but there are no targets in out-of-sample (test) data to evaluate the performance of such models. The test data set, which is used to evaluate generalization of the model predicting target t, can be also used to evaluate the performance of the PIM.

2702

Since the target values of PIs are not present in the test data set, the performance of the PIM can not be measured using classical residuals based (e.g. root mean squared error) measure. However, we can split the training data used to predict the target t, into two parts. The first part will be used to train PIM and the second part - to evaluate the generalization. It is worthwhile to point out that such split is not always possible when there is scarce of data. Even though target values of PIs are not present in the test data set, it is interesting to know whether target values t are inside the estimated prediction bounds (limits). Since, by definition, the prediction bounds enclose the true but unknown value ( 1 − α )% of times on average (typically 95%), we assess PIs estimation performance by evaluating the Prediction Interval Coverage Probability (PICP). The PICP is the probability that the target of an input pattern lies within the PIs and is estimated by the corresponding frequency as follows: 1 count (i ) V i : PLL ≤ t (i ) ≤ PLU

(8)

PICP =

Yet another performance measure for PIs can be introduced as well. This is the Mean Prediction Interval (MPI), which is an average prediction interval on test data set and it measures the ability to enclose target values inside the prediction bounds. MPI can be estimated by

[

1 V ∑ PI L (i ) + PIU (i ) V i =1

]

(9)

V. EXPERIMENTS & RESULTS A. Data Sets The method was tested on both artificial and real hydrologic data sets. The artificial data is generated using bivariate input variables x uniformly distributed between [0,1]. The true targets tt are given by: t t = f t ( x)

(10)

More specifically in our case, tt = b1 x1 + b2 x2

(11)

where b1 and b2 are linear regression coefficients and arbitrary values of these coefficients: 12 and 7 were used in the experiments. Since the noise is inherent to any real data set, we also added additive noise ε to tt to obtain new target t and it is estimated by the function f f as: t = t t + ε = f f ( x)

SNR = σ t / σ ε

(13)

where σ t and σ ε are the standard deviations of the target t and noise ε respectively. We varied the noise level by changing SNR values to 1, 3, 5 and 7. Total of 1500 instances were generated, and two-thirds of these selected randomly constituted the training data set, and the rest – the test data set. The hydrologic data set used in this experiment related to the river flows prediction in the Sieve catchment in Italy [4; 5]. Prediction of river flows Qt+i several hours ahead (i=1, 3 or 6) is based on using the previous values of flow (Qt − τq ) and previous values of rainfall ( REt − τr ) , where τq being between 0 and 2 hours and τr being between 0 and 5 hours. The sought regression models were based on 1854 examples and were formulated as follows:

{

where V is number of data in the test set. If the clustering technique and the PIM are optimal, then the PICP value will be consistently close to the ( 1 − α )%.

MPI =

The noise ε has Gaussian distribution N (0, σ t / SNR ) . SNR is signal to noise ratio defined by:

(12)

}

( j) nq ( j ) (14) SieveQi model → Qt + i = f f  {REt − τr }nr τr = 0 , Qt − τq τq = 0  i : {1,3,6}, j : {1,2,3}, nr : {5,3,0}, nq : {2,1,0}

Test data consisted of 300 instances. B. Procedure For a given data set, the main model f f was fitted on the training data set to estimate the target t. Then the trained model was used to predict the targets in the test data set. It is also possible that method can applied directly to the model outputs without running the forecasting model given that all or part of input variables (most influencing variables) to the forecasting model are available. Having the model outputs on the test data set, regression model PIM was constructed to estimate the PIs on test data set as follows. Fuzzy Cmeans clustering technique was first employed to construct the PIs for each cluster in the training data set and then to construct the PIs for each instance in the training data set. Note that the input to the PIM may constitute part or all of input variables, which are used in the forecasting model. The targets for the PIM are the upper and lower prediction bounds. Alternately, the target may be the upper and the lower PIs, but in this case, predicted values from the forecasting model have to be added to obtain the prediction bounds. In principle, structure or even class of the forecasting model and the PIM can be different (for example we may use neural networks for the forecasting model and linear regression for the PIM). For the purpose of comparison to other methods, in this paper we employed the same class of regression model. The PIs were constructed for 95% confidence level unless specified. The performance of the PIM is assessed by the PICM and the MPI introduced in Section IV.

2703

For the artificial data set, linear regression was fitted to predict both target t values and the PIs using least square methods. The optimal number of clusters in Fuzzy C-means clustering was identified using the performance index P [8] and the Xie-Beni index S [14]. Since the minimal values of these indices are not so pronounced, the experiments were repeated with the different numbers of clusters. For hydrologic data set, first bivariate linear regression was used as regression model using only two most influencing input variables (variables with higher correlation with output). Then later input variables were extended according to eq. (12). To estimate the effects of models complexity on the PIs, experiments were also conducted using MT and LWR as regression models.

contains input examples with very high runoff, whereas cluster 5 (C5) is associated with very low values of runoff. The two performance indicators: the PICP and the MPI were calculated for the PIM and the LRVE approach, and are visualized for artificial data set in Fig. 3. It can be seen that the PIM appears to perform better for data with high noise if compared to the LRVE approach. For lower noise the PICPs of the PIM are close to ( 1 − α )%. However, The PICP for the LRVE approach considerably deviate from the desired 95% confidence level. For all noise levels, the MPIs for the LRVE approach are wider than those of the PIM. Note that uncertainty increases as noise increases. 350

C. Results 800

250

600

3

Runoff (m /s)

C1 C2 C3 C4 C5

700

500

200

150

100

3

Runoff (m /s)

95% prediction bounds Forecasted runoff Observed runoff out of bounds

300

400

50

300

0

50

100

150

200

250

300

Time(hr)

200

Fig. 4. Computed prediction bounds for SieveQ1 100 0

0

500

1000 Time (hr)

100

1500

SieveQ1 SieveQ3 SieveQ6 ideal

80 PICP (%)

Fig. 2. Clustering of input examples in training data set in SieveQ1 using fuzzy C-means clustering 16 14

60 40 20

3

MPI (m /s)

12

0 10

0

8

40

60

80

100

100(1-alpha)% confidence level

Fig. 5. The PICP for different values of confidence level using MT as regression model and the PIM

6 4 2 92.5

20

93

93.5

94

94.5

95

95.5

96

96.5

PICP %

Fig. 3. Comparison of performance on artificial data set. Black blocks refer to the PIM, the grey blocks – to the LRVE approach. Bigger size of blocks are associated with higher noise (SNR=1), small circles are associated with low noise (SNR=7)

Fig. 2 shows clustering of input examples in Sieve catchment for 1 hour ahead prediction. Cluster 1 (C1)

The performance of the PIM is compared to that of the LRVE approach in case of the hydrologic data set with the bivariate input variable. It is observed that the PIM shows superior performance for all time leads of forecasts with respect to both the PICP and the MPI. The comparison of performance (in terms of the PICP and the MPI) for the hydrological data set (in this case more lagged input variables were used) using linear regression, LWR and MT both for the forecasting models and the PIM shows that performance of MT is better than that of the other models. However, the performance of linear

2704

regression and LWR are comparable. Fig. 4 shows the computed prediction bounds for SieveQ1 using MT. Fig. 5 shows the deviation of the PICPs from the desired confidence level for the hydrologic data set using MT. The PIs were constructed for various confidence levels ranging from 10% to 99%. It is to be noticed that the PICPs are very close to the desired confidence levels at values higher than 80% and in practice the PIs are constructed around this value. Furthermore, it can be noted that the PIs are too narrow in most of cases as the PICPs are below the straight line. Such evidence was also reported in [1]. In these cases the PIM underestimates uncertainty of the model outputs. Fig. 6 presents fan chart showing the MPIs for hydrologic data set with the different forecast lead times and different degrees of confidence levels. It is evident that the width of the PIs increases with the increase of the confidence level. Moreover it is also illustrated that the width of PIs increases as forecast lead time increases. Thus uncertainty of model forecast increases as lead time increases. 200

assumptions about the distribution of errors. LRVE and resampling approaches compute upper and lower prediction intervals symmetrically. The presented methodology computes upper and lower intervals independently. The methodology was applied to both artificial and real hydrologic data set, and was compared to LRVE approach typically used with the linear regression models; the superiority of the new method was demonstrated. We also demonstrate that the new method performs consistently better than the uniform interval method (UIM). The next step of this research is to compare the performance of our methodology to other techniques based on resampling methods ACKNOWLEDGMENT The authors acknowledge the valuable input of Mr. Javier Rodriguez during his Masters. study at UNESCO-IHE Institute for Water Education in 2003. This work is partly supported by the EU project “Integrated Flood Risk Analysis and Management Methodologies” (FLOODsite), contract GOCE-CT-2004-505420.

180

REFERENCES

160

[1]

120

[2]

3

MPI (m /s)

140

100 80

[3]

60

[4]

40 20 0

1

3 Forecast horizon (hour)

[5]

0 6

[6]

Fig. 6. A fan chart showing mean prediction intervals for flow prediction in Sieve catchment up to 6 hours ahead. The darkest strip covers 10% probability and the lightest - 99%.

[7]

We also compared the results with the Uniform Interval Method (UIM) that constructs single PI from empirical distribution of errors on the whole training data and applied uniformly to the test data set. The PIM performs consistently better than the UIM both in terms of the PICP and the MPI. Due to lack of space, the results are not presented here. VI. CONCLUSIONS A novel method to compute the prediction intervals (PIs) for the model outputs using machine learning is presented. Computed PIs explicitly take into account all sources of uncertainty of the model outputs. The methodology is independent of the structure of the forecasting model as it requires only the model outputs. Unlike existing techniques the methodology does not require the knowledge of prior distribution of parameters and does not rely on any

[8] [9] [10] [11] [12] [13] [14]

2705

C. Chatfield, “Time-Series Forecasting”, Chapman & Hall/CRC, 2000. D. L. Harnett, and J. L. Murphy,” Introductory Statistical Analysis”, Addison-Wesley publishing compay, 1980. D. Nix, and A. Weigend, “Estimating the mean and variance of the target probability distribution”, Proc. of the IJCNN, IEEE, pp. 55-60, 1994. D. P. Solomatine, and D. L Shrestha, “AdaBoost.RT: a Boosting Algorithm for Regression Problems”, Proc. of the Int. Joint Conference on Neural Network, Budapest, 2004. D. P. Solomatine, and K. N. Dulal, “Model tree as an alternative to neural network in rainfall-runoff modelling”, Hydrological Science Journal, vol. 48(3), pp. 399–411, 2003. J. C. Bezdek, “Pattern Recognition with Fuzzy Objective Function Algorithms”, Kluwer Academic Publishers Norwell, MA, USA, 1981. K. J. Beven, and J. Binely, “The future of distributed models: model calibration and uncertainty prediction”, Hydrological Processes, vol. 6, pp. 279-298, 1992. M. Amiri, “Yashil's Fuzzy C-Means Clustering MATLAB Toolbox”, Ver. 1.0. Available: http://yashil.20m.com/. R. Krzysztofowicz, “The case for probabilistic forecasting in hydrology”, J. of Hydrology, vol. 249, pp. 2-9, 2000. S. Geman, E. Bienenstock, and R. Doursat, “Neural networks and the bias/variance dilemma”, Neural Computation, vol. 4, pp. 1-58, 1992. T. H. Wonnacott, and R. J. Wonnacott, “Introductory Statistics”, John Wiley & Sons, Inc, 1996. T. Heskes, “Practical confidence and prediction intervals”, Advances in Neural Information Processing Systems 9, M. Mozer et al. (eds.), MIT Press, pp. 176-182, 1997. W. D. Penny, and S. J. Roberts, “Neural network predictions with error bars”, Research report TR-97-1, Dep. of Electrical and Electronic Engg., Imperial College, London, 1997. X. L. Xie, and G. Beni, “A Validity Measure for fuzzy clustering,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 13, no. 8, pp. 841-847, 1991.

Estimation of Prediction Intervals for the Model Outputs ...

Abstract− A new method for estimating prediction intervals for a model output using machine learning is presented. In it, first the prediction intervals for in-.

230KB Sizes 1 Downloads 270 Views

Recommend Documents

A Hybrid Prediction Model for Moving Objects - University of Queensland
for a data mining process. ... measures. • We present a novel data access method, Trajectory Pat- ..... node has free space, pk is inserted into it, otherwise it splits.

A Hybrid Prediction Model for Moving Objects - University of Queensland
a shopping center currently (9:05 a.m.), it is unreasonable to predict what her ..... calls, hence, the computing time is significantly reduced. ..... an eight hour period. .... 24. 26. 28. 30. 32. 34. 36. 38. Eps. ( a ) number of patte rn s. Bike. C

An Improved Particle Swarm Optimization for Prediction Model of ...
An Improved Particle Swarm Optimization for Prediction Model of. Macromolecular Structure. Fuli RONG, Yang YI,Yang HU. Information Science School ...

An Estimation Model for the Savings Achievable by ...
tool chain. Keywords-Cost Estimation; Tool Integration; Software Pro- cess Improvement ... is the integration of tools from different tool vendors and across different ... historical project data. Most algorithmic ..... B. Steece, “COCOMO II Model

Maximum likelihood estimation of the multivariate normal mixture model
multivariate normal mixture model. ∗. Otilia Boldea. Jan R. Magnus. May 2008. Revision accepted May 15, 2009. Forthcoming in: Journal of the American ...

Semiparametric Estimation of the Random Utility Model ...
Apr 15, 2017 - ... the consistent estimation of the ratios of coefficients despite stochastic mis- ... is asymptotically normal, meaning that it is amenable to the ...

A Three-dimensional Dynamic Posture Prediction Model for ...
A three-dimensional dynamic posture prediction model for simulating in-vehicle seated reaching movements is presented. The model employs a four-segment ...

System and method for synchronization of video display outputs from ...
Jun 16, 2009 - by executing an interrupt service routine by all host processors. FIG. 9 .... storage medium or a computer netWork Wherein program instructions are sent over ..... other information include initialization information such as a.

A Self-Similar Traffic Prediction Model for Dynamic ...
known about the traffic characteristics of wireless networks. It was shown in [1] that wireless traffic traces do indeed exhibit a certain degree of self-similarity and ...

Collision model for vehicle motion prediction after light ...
Vehicle collision mechanics are useful in fields such as vehicle crashworthiness, passenger injury prediction, and ... the prediction results of this 4-DOF model against the commercial vehicle dynamics software. CarSimTM. The rest ... Section 3 prese

A Self-Similar Traffic Prediction Model for Dynamic ...
The availability of precise high-quality and high-volume data sets of traffic ... to forecast real-time traffic workload could make dynamic resource allocation more ...

A Hierarchical Model for Value Estimation in Sponsored ...
Also, our hierarchical model achieves the best fit for utilities that are close to .... free equilibrium, while Varian [22] showed how it could be applied to derive bounds on ...... the 18th International World Wide Web Conference, pages 511–520 ..

two-dimensional cfd model for the prediction of flow ...
of superficial gas velocity (VG), column diameter (D), column height (HD) and ... and heat transfer coefficient were in good agreement with the experimental data.

Empirical calibration of confidence intervals - GitHub
Jun 29, 2017 - 5 Confidence interval calibration. 6. 5.1 Calibrating the confidence interval . ... estimate for GI bleed for dabigatran versus warfarin is 0.7, with a very tight CI. ... 6. 8. 10. True effect size. Systematic error mean (plus and min.

Robust Nonparametric Confidence Intervals for ...
results are readily available in R and STATA using our companion software ..... mance in finite samples by accounting for the added variability introduced by.

Soft Margin Estimation of Hidden Markov Model ...
measured by the generalization ability of the machine learning algorithms. In particular, large margin classification tools, such as support vector machines ...

System and method for synchronization of video display outputs from ...
Jun 16, 2009 - media include magnetic media such as hard disks, ?oppy disks, and ... encompass data signals embodied in a carrier Wave such as the data ...

Particle Filters for the Estimation of a State Space Model
monitoring and process control. State estimation can be ..... Gaussian noise. Thus the objective of the state estimation task is to infer CA and T sequentially given ...

Structural estimation of sovereign default model
We quantitatively evaluate DSGE model of the emerging economy with .... Flury and Shephard (2011) and Malik and Pitt (2011) estimate simple DSGE model with particle filter. ..... News and sovereign default risk in small open economies.

Incentives for Eliciting Confidence Intervals
Dec 3, 2012 - rule punishes the expert for specifying a larger interval width, and for the distance of ..... In sum, learning about the dispersion of beliefs is confounded. ... width gives information about their relative degrees of risk aversion.

Two-Sided Confidence Intervals for the Single Proportion
Jun 9, 2017 - See discussions, stats, and author profiles for this publication at: ... zero width intervals; and ease of use, whether by tables, software or formulae. .... variance p(1!p)/n, and we proceed to show that this results in a good degree .

Comparing Machine Learning Methods in Estimation of Model ...
Abstract− The paper presents a generalization of the framework for assessment of predictive models uncertainty using machine learning techniques. Historical ...