2006 International Joint Conference on Neural Networks Sheraton Vancouver Wall Centre Hotel, Vancouver, BC, Canada July 16-21, 2006

Eager and Lazy Learning Methods in the Context of Hydrologic Forecasting Dimitri. P. Solomatine, Member, IEEE, Mahesh Maskey, and Durga Lal Shrestha, Member, IEEE 1

Abstract− Computational intelligence techniques are becoming popular in hydrologic forecasting. Primarily these are eager learning methods. Lazy (instance-based) learning (IBL) has received relatively little attention, and the present paper explores the applicability of these methods. Their performance is compared with that of neural networks, M5 model trees, regression trees. A flow forecasting problem was solved along with the five benchmark problems. Results showed that one of the IBL methods, the locally weighted regression, especially if used with the Gaussian kernel function, often is more accurate than the eager learning methods.

I. INTRODUCTION Accurate forecastings of environmental variables like precipitation, runoff, water stages, etc. are of major focus in hydrologic modeling. Apart from the methods based on the detailed description of the physical processes (process or physically-based), the methods based on computational intelligence (CI) (often referred to as data-driven models) are gaining popularity [26]. In the context of hydrologic modeling, data driven models are typically based on the historical records about the relevant input (e.g., rainfall and temperature) and output (flow) variables, and they make a limited number of assumptions about the details of the processes transforming the rainfall into runoff. Among the various types of data driven models, an artificial neural network (ANN) is the most popular choice – see e.g., [13], [16], [8], [1], [12]. Along with ANN, other numerical prediction (regression) methods are used as well: Solomatine and Dulal [24] applied the so-called M5 model trees (MT); Bray and Han [5] used support vector machines; Solomatine and Xue [25] used the modular models (committees) comprised of ANNs and M5 model trees [18]. ANN is accurate in reconstructing complex non-linear dependencies but suffers from a problem of being encapsulated in software codes and therefore not transparent enough, which is an issue during their acceptance in hydrology. Dimitri P. Solomatine is with the UNESCO-IHE Institute for Water Education, P.O. Box 3015, 2601 DA Delft, The Netherlands (corresponding author, phone: +31-15-2151815, e-mail: [email protected]) Mahesh Maskey is with NepalConsult (P.) Ltd., P.O. Box 92, Gushingal, Lalitpur, Kathmandu, Nepal (email:[email protected]) Durga Lal Shrestha is with the UNESCO-IHE Institute for Water Education, P.O. Box 3015, 2601 DA Delft, The Netherlands (e-mail: [email protected])

0-7803-9490-9/06/$20.00/©2006 IEEE

One of the techniques in machine learning that has a potential to resolve the issue of non-transparency is instancebased learning when prediction is made on the basis of combining historical examples (instances) that are in some way close to the new vector of inputs. In hydrology one can find quite a very limited number of examples of using this class of methods. Karlsson and Yakowitz [14] were probably the first to use this method in hydrology, focusing, however only on (single-variate) time series forecasts by the k-nearest neighbor method (k-NN). Given time series {xt}, t=1,…, n, they generated d-dimensional vectors xd(t) = { xt , xt-1 , …, xtd+1 } (for t=d,…,n-1) and based the prediction of xT+1 on averaging the values xt+1 that corresponded to the k vectors in this d-dimensional space that are close to xd(T). (Interestingly, their approach has an intuitive relation to the single-variate predictors based on non-linear dynamics and chaos theory, which however provides a much more solid foundation for this type of prediction.) Galeati [10] demonstrated the applicability of the k-NN method (with the vectors composed of the lagged rainfall and flow values) for daily discharge forecasting and it compared favorably to the ARX statistical model. Shamseldin and O’Connor [21] used the k-NN method in adjusting the parameters of the linear perturbation model for river flow forecasting. Toth et al. [28] compared the k-NN approach to other time series prediction methods in a problem of short-term rainfall forecasting. In the present paper instance-based learning is considered in a wider context of machine learning, several methods are explored, their applicability in hydrologic forecasting is tested and their performance is compared to other methods. II. DATA-DRIVEN MODELS IN HYDROLOGY In this paper by a data-driven model (DDM) the following model will be understood: y = f (X)

(1)

where f = machine learning (e.g., ANN) or statistical (e.g., linear regression) model whose internal parameters are found by calibration (i.e., training, or optimization); y = scalar (typically, real-valued) output; X ∈ Rn (n-dimensional real-valued input vector). The training set is denoted as T, and the verification (test) set, as V. As the measure of model error root mean square error (RMSE) or so-called volumetric fit are used.

9901

For an example of a DDM that can be used for hydrologic forecasting, we may turn to the paper by Solomatine and Dulal [24] where several machine learning models were built to predict the river flows Qt+i several hours ahead (prediction horizon H=1, 3 or 6): Qt+H = f (REt-τr, Qt-τq)

most common value among k training examples nearest to q. For real valued target functions, the estimate is the mean value of the k-nearest neighboring examples. The k-NN algorithm can be improved by weighting each of the kneighbors Xi according to their distance to the query point q so that the output value for q is calculated as follows:

(2)

where Qt-τq = previous (lagged) values of flow; REt-τr = previous values of rainfall; τq ∈ [0, 2] hours; τr ∈ [0, 5] hours (see also (8)). The values of lags τq and τr are based on the hydrologic analysis of the catchment, and on the analysis of correlation and average mutual information between inputs and outputs. In the notation of model (1) dimension n of the input vector X is equal to the total number of the past rainfall and flow values used as inputs. For example, for the prediction horizon H=3 vector X = {REt, REt-1, REt-2, REt-3, Qt-1, Qt}, so n = 6. The training set T is composed of all past records of the properly lagged values of rainfall and flow arranged as 6-dimensional vectors, each accompanied by the value of the measured value of flow H time steps ahead. III. INSTANCE-BASED LEARNING (IBL) CI methods following the eager learning paradigm, construct a general explicit description of the target function when training examples are provided. In contrast, instancebased learning (IBL) is referred to as lazy learning since it is seen as consisting of simply storing the presented training data. Then when a new input vector is presented, a set of similar related instances is retrieved from memory and their corresponding outputs are used to predict the output for the new query vector (instance). IBL algorithms are derived from the nearest neighbor pattern classifier [7], [2], [17]. In IBL function f in (1) is, in fact, never explicitly built. These methods construct a local approximation to the modeled function (1) that applies in the neighborhood of the new query instance encountered. Thus it describes the very complex target function as a collection of less complex local approximations. IBL algorithms have several advantages: they are quite simple but robust learning algorithms, can tolerate noise and irrelevant attributes, and can represent both probabilistic and overlapping concepts and naturally exploit inter-attribute relationships [2]. IBL can be time-consuming requiring O (|T|×n) attribute examinations, where T is the training set and n is the dimension of input space. A. k-Nearest Neighbors and variations The nearest neighbor classifier is one of the simplest and oldest methods for classification. It classifies an unknown input vector Xq (denoted further also as q) by choosing the class of the nearest example X in the training set as measured by a distance metric, typically Euclidean. Generalization of this method is the k-nearest neighbor (k-NN) method. For a discrete valued target function, the estimate will just be the

k

k

i =1

i =1

f ( q ) = ∑ wi f ( X i ) / ∑ wi

(3)

where weight wi is a function of the distance d(Xq, Xi) between Xq and Xi. Typically, the following weight functions are used: (a) wi = 1 – d(Xq, Xi); (c) wi = d(Xq, Xi ))-2.

(b) wi = (d(Xq, Xi))-1 ;

(4)

In Weka software (Witten and Frank, 2000) used in this research, the functions (a) and (b) are implemented. B. Locally Weighted Regression (LWR) Locally weighted regression (LWR) is inspired by the instance-based methods for classification [3]. In it, the regression model is built only when the output value for a new vector q should be predicted, so that all learning is performed at prediction time. It uses linear or non-linear regression to fit models locally to particular areas of instance space in a way quite different from M5 model trees. The training instances are assigned weights wi according to their distance to the query instance q and regression equations are generated on the weighted data. A number of distance-based weighting schemes can be used in LWR [20]. A common choice is to compute the weight wi of each instance Xi according to the inverse of their Euclidean distance d (Xi, Xq) from the query instance q wi = K (d ( X i , X q )) = 1/ d ( X i , X q )

(5)

where K (.) is typically referred to as the kernel function, and d (.) is the distance function. Atkeson et al. [3] combined Euclidean distance with the Gaussian kernel function: wi = K (d ( X i , X q )) = exp(−d ( X i , X q )2

(6)

Alternatively, instead of weighting the data directly, the model errors for each instance used in the regression equation are weighted to form the total error criterion C(q) to be minimized:

9902

T

C (q) = ∑ L ( f ( X i , β ) , yi ) K (d ( X i , X q )) i =1

(7)

where f (Xi, β) = regression model giving an output value estimate yi* ; L(yi*, yi ) = error function (typically the sum of squared differences (yi * – yi )2 between the target yi* and estimated yi output values); β is a vector of parameters to be identified; yi = target output value for the input vector Xi ; i = 1…|T |; T = training set. Gasser and Muller [11], Cleveland and Loader [6] and Fedorov et al. [9] address the issue of choosing weighting (kernel) functions: it should be maximum at zero distance, and the function should decay smoothly as the distance increases. Discontinuities in the weighting functions lead to discontinuities in the predictions, since training points cross the discontinuity as the query changes. Yet another possibility to improve the accuracy of LWR is to use the smoothing, or bandwidth parameter that scales the distance function by dividing it by this parameter [20], [6]. One way to choose it is to set it to the distance to the kth nearest training instance, so that its value becomes smaller as the volume of training data increases. An appropriate smoothing parameter can be found using cross-validation. One can see certain analogy between LWR and the radialbasis function ANNs.

used in IBL predictor. Such an approach is quite general and may involve any IBL method and any predictive model. V. CASE STUDIES In the present study a problems of hydrologic forecasting for Bagmati catchment was considered. Additionally, the applied methods were tested on the five standard machine learning benchmark data sets. A. Bagmati catchment Bagmati catchment lies in the central part of Nepal (Fig.1). It is a medium sized foothill fed river basin with an area of about 3700 km2 and originates from the southern slope of Shivapuri lake (Mahabharat within Kathmandu valley) and stretches to the plains of Terai (ending at NepalIndia border). The catchment covers eight districts of Nepal and is a perennial water body of Kathmandu. The problem was posed as a short-term flow forecasting at Pandheradobhan hydrometric station.

IV. COMBINING MODELS: COMMITTEES AND COMPOSITE MODELS

Combination of classification or regression models often brings improvements in accuracy (e.g. [31], [29], and [15]). Solomatine and Siek [27] distinguish between (1) modular models when separate models are trained on different subsets of input data; (2) committees (ensembles) of models when they are trained on the same dataset and the results are combined by some “voting” scheme, and (3) complementary (composite) models when one is used to correct errors of another. Solomatine and Xue [25] used the first approach (mixtures of models) in the flow predictions in the Huai river basin (China). In this study we use the second approach where a committee combines instance-based models, M5 model trees and neural networks, and the third approach where a M5 model tree is complemented by an instancebased model. One of the methods combining various models is that of Quinlan [19] – it combines IBL with M5 model trees with and is further referred as “composite model”. Such approach is (supposedly) implemented in the Cubist software [8], and is outlined below. For an unseen example q, the target value y is to be predicted. A subset of vectors (prototypes) {X1, X2,.., Xk} would first be identified as nearest to q. In a standard IBL method the known values {f (X1), f (X2),...., f(Xk)} would be combined to give the predicted value of the unseen example q. In the composite model, however, these values are adjusted in the following way. Among such prototypes, one can be selected, say Xi. Now some model (Quinlan suggests M5 model tree) is used to predict target values so that its predictions for q and Xi are f*(q) and f*(Xi ) respectively. Instead of f(Xi), the adjusted value f(Xi) – (f*(Xi ) – f*(q)) is

Fig.1. Bagmati catchment (triangles denote the rainfall stations; the flow is measured at Pandheradobhan)

Rainfall is measured at the three stations within the basin with the daily resolution for eight years (1988 to 1995) was collected, however, only the mean areal rainfall was available for this study. The computed daily evapotranspiration was subtracted from the rainfall and the resulting effective rainfall was used in this study. Analysis of the relationships between the input and output variables was done by visual inspection and the correlation analysis. The lag of one day was accepted as the average lag time of rainfall (in (2) this is τr). The forecasting model had five input variables used to predict the flow one time step ahead: Qt+1=f (REt-2, REt-1, REt, Qt-1, Qt)

(8)

where Q=discharge; RE=effective rainfall (mean rainfall minus evapotranspiration).

9903

In hydrologic context, splitting the data into training and testing data sets in hydrology may not be easy. Ideally, these sets should include approximately equal number of precipitation events, have similar distribution of the low and high flows, or, in other words, the input and output variables should be statistically similar – have similar distributions, or at least mean, variance and range. However, a constraint typical to hydrologic studies is that the test data should be a set of points contiguous in time – this makes the generation of the training and test sets with the similar statistical properties not an easy task and leaves not too many choices. For the present study the eight years of data (2919 fivedimensional vector instances) were split as follows: the first 919 vectors (3-Jan-1988 to 7-Jul-1990) were used as the testing data set while the remaining vectors (8-Jul-1990 to 30-Dec-1995) were used for training. Table I presents the statistical properties of the mentioned sets.

For distance-weighted k-nearest neighbors (further abbreviated as IBk) implemented in Weka software [30], the first two weight functions (4) were used. For locally weighted regression (LWR) implemented in Weka, all three types of kernel functions, linear weighting, inverse weighting and Gaussian kernel functions, were used. Cubist software (http:// www.rulequest.com) was used as well. It builds two types of models: (1) rules, which are seemingly generated from M5 model trees and will be referred to as MT(C) (this is a proprietary method and no details of how exactly such rules are generated could not be found in the literature), and (2) a “composite model” [19] combining rules from model tree and k-NN method which will be referred as MT(C)+k-NN. Besides, for comparison the model tree (MT) and regression tree (RT) models developed by Solomatine and Dulal [24] were used.

TABLE I STATISTICAL PROPERTIES OF THE BAGMATI DATA SETS

Bagmati catchment. For this catchment we also included the results reported by Shrestha [22] on using ANN. Two data sets were used to predict flow Qt+1 one day ahead with the same sets of input attributes. The first experiment involved the original data set. Further, the same data set was randomized to allow for generation of training and test sets that are statistically similar. Table III presents the results. The number of nearest neighbors for IBL methods shown was found from crossvalidation. In both versions of data split, LWR had the best performance. For the original data set, IBk is better than MT; it is opposite for the randomized data sets. IBL methods have higher accuracy than the ANN built for Bagmati original data set.

Discharge Average Minimum Maxi Std. dev. Skewness

All data 149.96 5.10 5030 271.12 6.45

Training set 160.84 5.10 5030 291.42 6.79

Test set 126.24 8.20 2110 218.62 3.83

Note: Std. dev. = standard deviation. Apart from that, a procedure was developed to perform a proper split into training and test sets was also made that ensured their close statistical proximity, leaving however the records not consecutive in time.

B. Results

TABLE III COMPARISON OF MODEL PERFORMANCE IN TERMS OF RMSE FOR BAGMATI

TABLE II CHARACTERISATION OF THE FIVE BENCHMARK DATA SETS

Data sets Autompg Bodyfat CPU Friedman Housing

Instances 398 252 209 1500 506

Attributes 9 15 8 6 14

DATA SETS

Attributes types 3 discrete, 5 real-valued All real-valued All real-valued All real-valued 1 binary-valued

Methods

B. Benchmark data sets In the experiments with the machine learning methods, it is important to provide a reference to the standard benchmark sets typically used to test such methods. The five data sets from the popular UCI Repository [4] were considered: Autompg, Bodyfat, CPU, Friedman and Housing (Table II). VI. EXPERIMENTS AND RESULTS A. Models set up For IBL algorithms one of the main parameters is the number of nearest neighbors. In the present study, 1, 3, 5 and 9 nearest neighbors were considered.

Original

Randomized

Training

Testing

Training

Testing

ANN

100.1

163.1

-

-

MT

97.70

113.2

103.9

142.4

RT

204.3

153.7

121.4

156.4

LWR

87.74

107.0

82.5

127.3

IBk

35.50

107.8

32.7

148.3

MT(C)

101.8

115.5

99.85

139.2

Comp

108.7

108.3

102.9

135.8

Note: Comp = MT(C)+k-NN. LWR and IBk use 9 numbers of neighbors; MT(C)+k-NN uses 5 numbers of neighbors.

Fig. 2 shows the comparison of various instance-based learners. The accuracy of the three models is similar, except for some points. For high flows errors are quite high. In order to give an idea of the structure of rules generated by Cubist on the basis of the built M5 model tree (piece-wise linear model) the three rules (out of 13) are presented below.

9904

2000 1800

Discharge(m3/s)

1600 1400 1200 1000 800 600 400 200 0 500

520

540

560

580

600

620

640

660

680

700

Time (days) Observed Values

LWR 9 NN Gaussian

IBk 9 NN Inverse Distance

MT (C) + 5 NN

Fig. 2. Comparison of various instance-based learners in testing: LWR, IBk and MT(C)+k-NN (with 5 numbers of neighbors) on Bagmati data set (fragment with the instances 500..700, i.e. from 14-May-89 to 3-Nov-89)

The algorithm splits the data set into subsets (the number of examples is given in parentheses) and builds linear regression models for each of them. In fact the splitting into such subsets often has a reasonable hydrologic interpretation of representing various types of hydrologic conditions – low flows, high past precipitation and low current flow, high flows, etc.

TABLE IV COMPARISON OF VARIOUS MODELS FOR BENCHMARK DATA SETS

Data set Autompg Bodyfat

1. if Qt <= 40.1 then Qt+1 = 1.132 + 0.97 Qt + 1.3 REt (935 examples) 2. if REt <= 20.099 & Qt > 40.1 & Qt <= 222 then Qt+1 = 3.233 + 0.99 Qt + 2.1 REt (553 examples) 3. if REt <= 39.017 & Qt > 222 & Qt <= 297 then Qt+1 = 70.046 + 0.65 Qt + 2 REt + 0.03 Qt-1 (137 examples)

CPU Friedman Housing

Benchmark data sets. Table IV shows the comparison of various models for all data sets. (Note that the errors in training for a lazy learning method may not be an appropriate measure of its performance: e.g., 1-nearest neighbor method will always have zero training error.) LWR worked well in all cases; on the Bodyfat data set all IBL methods have shown almost negligible errors. On all data sets the composite model (MT(C)+k-NN) showed better performance than the non-IBL MT(C). Finally, the scoring matrix [23] was used to determine which method is the best overall (on the test set). This is a square matrix in which the elements SMi,j are the average of the relative performance of the ith algorithm over the jth algorithm with respect to all data sets used (it can be negative); diagonal elements are zero:

MT

RT

LWR

IBk

Train 2.50

2.25

2.10

0.41

2.86

2.77

Test

2.46

3.14

2.23

2.89

2.60

2.42

Train 0.67

0.71

0.00

0.00

0.62

0.63

Test

0.39

0.00

0.00

0.22

0.00

Train 28.48 32.57

4.16

0.00

25.14

26.95

0.39

MT(C) Comp

Test 43.44 45.68 32.92 56.64

40.34

37.91

Train 2.21

2.20

0.54

0.07

1.58

1.03

Test

2.20

2.72

1.38

1.79

1.80

1.56

Train 2.82

3.25

2.11

0.44

2.56

2.40

Test

2.83

1.92

2.74

2.27

2.01

2.25

Note: Comp = MT(C)+k-NN. LWR and IBk use 9 numbers of neighbors; MT(C)+k-NN uses 5 numbers of neighbors. Bold type face signifies the minimum value of RMSE for each test data set.

RMSEk , j − RMSEk ,i 1 N ,  ∑ SM i , j =  N k =1 max( RMSEk , j , RMSEk ,i )  0, i = j

i≠ j

(9) where N is the number of data sets. By summing up all matrix values column-wise one can determine the overall score of each algorithm. Table V shows the scoring matrix for all algorithms and data sets. The composite model MT(C)+k-NN was found to be the best (scoring factor of 87.97%). LWR was the best in

9905

most of the case studies, but its superiority was not as significant as that of MT(C)+k-NN on other cases, so overall it was the second best (55.25%).

the new data vector. IBL approach makes the prediction transparent and explainable. VIII. CONCLUSIONS

Table V SCORING MATRIX SHOWING OVERALL PERFORMANCE OF THE METHODS

Models MT RT

MT

RT

LWR

IBk MT(C) Comp Score Rank

0.00 20.47 -15.03 1.26

-7.09 -17.45 -17.85

5

-20.47 0.00 -33.67 -18.43 -26.59 -36.33 -135.49 6

LWR

15.03 33.67 0.00

4.84

13.99 -12.29 55.25

2

IBk

-1.26 18.43 -4.84

0.00

-2.70

MT(C) 7.09 26.59 -13.99 2.70

-6.66

2.97

4

0.00 -15.24

7.15

3

Comp 17.45 36.33 12.29 6.66 15.24 0.00 87.97 1 Note: Comp = MT(C)+k-NN. All values are given in percentage.

VII. DISCUSSION The performed experiments showed that IBL methods are accurate predictors in seven data sets out of nine, and LWR was the winner. Concerning the choice of the method parameters, in LWR the Gaussian kernel function, and in IBk (k-NN method) the inverse weighted distance were the best choices. Is it possible to make a universal judgment about the appropriateness of this or that IBL method for hydrologic forecasting? We do not think so. Any machine learning (data-driven) method can excel on one data set and show a meager performance on another, and there are no universally applicable rules for selection of the method that would be best in all cases. For the presented experiments it can be said that the IBL methods and the M5 model trees perform the “local” modeling (modular models, committees and IBL), that is they are using models built (trained) on the basis of a subset of the whole data set. In the presented hydrologic problems these methods have superior performance if compared with the “global” methods where models are trained on the whole data set (ANNs and conceptual models). In hydrologic context this may mean that the modeled processes consist, in fact, of a number of different processes (e.g., resulting in low, medium and high flows) each of which should, in principle, be modeled separately. The essence of using IBL methods for hydrologic forecasting is in following a simple idea: use the flow value (or a function of several values) that resulted from the similar hydrologic situations in the past. In this respect it is of utmost importance to choose the relevant input variables which are properly lagged – this is where the knowledge of hydrology of the catchment comes into play and directly used in a DDM. It is also important to stress that one of the features of IBL is that it is possible for model users and decision makers to judge why and how a certain prediction is made. The reason is that it is not too difficult to find the nearest neighbors of

The performed experiments with the selected case studies showed that the IBL methods are accurate numerical prediction tools: they were more accurate than the other methods in seven data sets out of nine. Dependency of the model accuracy on the choice of various kernels for LWR and distance functions was investigated. It was also noted that the use of the inverse weighted kernel functions lead to the higher computational time than in the case of using linear and Gaussian kernels (linear weighting kernel function is the fastest but inaccurate). The performed experiments have also confirmed the higher accuracy of hybrid models, represented in this study by the composite model of Quinlan [19] combining a model tree (multi-linear model) and IBL. In classification of flow, the k-NN method showed high accuracy. When k-NN is used, different distance functions should be tried first and then the best one adopted. Overall, the IBL methods, especially local weighted regression, appear to be accurate numerical predictors and can be successfully used in forecasting. In the context of hydrologic modeling, IBL and modular models like M5 model trees can be seen as a combination of “local” models, each responsible for forecasting in a particular region of the input space – corresponding to a particular hydrologic condition. IBL, using the discharges resulting from the similar past hydrologic situations to compute the forecast, is dependent on the appropriate choice of the lagged hydrologic variables characterizing such conditions. The hydrologic characteristics of the catchment are embodied in the set of input variables and in this sense data-driven models cannot be considered “black-box” models. Instance-based learning methods, together with the M5 model trees [24], [27], can be seen as important alternatives to statistical models, non-linear methods like ANNs, and may play important role in hydrologic forecasting, complementing thus the physically-based distributed models. They also have an advantage of being more transparent than ANNs and hence may be easier accepted by decision makers. The further directions of research are seen in (1) extending the models types that are combined into a modular model, and (2) considering the interpretable hydrologic events together with the simulation model runs as inputs to instance-based learning. ACKNOWLEDGMENT This work is partly supported by the EU project “Integrated Flood Risk Analysis and Management Methodologies” (FLOODsite), contract GOCE-CT-2004505420.

9906

REFERENCES [1] R. J. Abrahart and L. See, “Comparing neural network and auto regressive moving average techniques for the provision of continuous river flow forecasts in two contrasting catchments,” Hydrological Processes, 14, pp. 2157–2172, 2000. [2] D. Aha, D. Kibler, and M. Albert, “Instance-based learning algorithms,” Machine Learning, Boston: Kluwer Academic Publishers, pp. 37-66, 1991. [3] C. G. Atkeson, A. W. Moore, and S. Schaal, “Locally weighted learning,” Artificial Intelligence Review, 11, pp. 1173, 1996. [4] C. L. Blake and C. J. Merz, “UCI repository of machine learning databases,” Irvine, CA: University of California, Dep. of Information and Computer Science, 1998. Available online at http://www.ics.uci.edu/~mlearn/MLRepository.html. [5] M. Bray and D. Han, “Identification of support vector machines for runoff modelling,” Journal of Hydroinformatics, 6, pp. 265-280, 2004. [6] W. S. Cleveland and C. Loader, “Smoothing by local regression: Principles and methods,” Technical Report 95.3. AT & T Bell Laboratories, Statistics Department, Murray Hill, NJ, 1994. [7] T. M. Cover and P. E. Hart, “Nearest neighbour pattern classification,” IEEE Transactions on Information Theory, 13, pp. 21-27, 1997. [8] Y. B. Dibike and D. P. Solomatine, “River flow forecasting using artificial neural network,” Journal of Physics, Chemistry of the Earth, Part B: Hydrology, Oceans and Atmosphere, 26(1), pp. 1–8, 2001. [9] V. V. Fedorov, P. Hackl, and W. G. Muller, “Moving local regression: The weight function,” Nonparametric Statistics, 2(4), pp. 355-368, 1993 [10] G. Galeati, “A comparison of parametric and non-parametric methods for runoff forecasting,” Hydrological Sciences Journal, 35(1), pp. 79-94, 1990. [11] T. Gasser and H. G. Muller, “Kernel Estimation of regression functions,” in Smoothing Techniques for Curve Estimation, Edited by T. Gasser and M. Rosenblatt, pp. 23-67, 1979. [12] R. S. Govindaraju and A. Ramachandra Rao, Artificial neural networks in hydrology, Kluwer Academic Publishers, 2001. [13] K. Hsu, H. V. Gupta, and S. Sorooshian, “Artificial neural network modelling of the rainfall-runoff process,” Water Resources Research, 31(10), pp. 2517-2530, 1995. [14] M. Karlsson and S. Yakowitz, “Nearest neighbour methods for non-parametric rainfall runoff forecasting,” Water Resources Research, 23(7), pp. 1300-1308, 1987. [15] L. L. Kuncheva, Combining Pattern Classifiers, New York: Wiley, 2004.

[16] A. W. Minns and M. J. Hall, “Artificial neural networks as rainfall-runoff models,” Hydrological Sciences Journal, 41(3), pp. 399-417, 1996. [17] T. Mitchell, Machine learning, MIT Press and McGraw-Hill, 1997. [18] J. R. Quinlan, “Learning with continuous classes,” Proc. AI’92, 5th Australian Joint Conference on Artificial Intelligence, Edited by Adams and Sterling, Singapore: World Scientific, pp. 343-348, 1992. [19] J. R. Quinlan, “Combining instance-based and model-based learning,” Proc. ML’93, Edited by P. E. Utgoff, San Mateo, CA: Morgan-Kaufmann, 1993. [20] D. W. Scott, Multivariate Density Estimation, New York: Wiley, 1992. [21] A. Y. Shamseldin and K. M. O’Connor, “A nearest neighbour linear perturbation model for river flow forecasting,” Journal of Hydrology, 179, pp. 353-375, 1996. [22] I. Shrestha, “Conceptual and data-driven hydrological modelling of Bagmati river basin, Nepal,” M. Sc. Thesis HH451, IHE Delft, The Netherlands, 2003. [23] D. L. Shrestha and D. P. Solomatine, “Experiments with AdaBoost.RT, an Improved Boosting Scheme for Regression,” Neural Computation, 18(4), 2006. [24] D. P. Solomatine, and K. N. Dulal, “Model tree as an alternative to neural network in rainfall-runoff modelling,” Hydrological Sciences Journal, 48(3), pp. 399-411, 2003. [25] D. P. Solomatine, and Y. Xue, “M5 model trees and neural networks: application to flood forecasting in the upper reach of the Huai River in China,” ASCE Journal of Hydrological Engineering 9(6), pp. 491-501, 2004. [26] D. P. Solomatine, “Data-driven modelling and computational intelligence methods in hydrology,” Encyclopaedia of Hydrological Sciences, Edited by M. Anderson, New York: Wiley, 2005. [27] D. P. Solomatine, M. B. Siek, “Modular Learning Models in Forecasting Natural Phenomena,” Neural Networks, 19(2), pp. 215-224, 2006. [28] E. Toth, A. Brath, and A. Montanari, Comparison of shortterm rainfall prediction models for real-time flood forecasting,” Journal of Hydrology, 239, pp. 132-147, 2000. [29] S. M. Weiss, and N. Indurkhya, “Rule-based machine learning methods for functional prediction,” Journal of Artificial Intelligence Research, 3, pp. 383-403, 1995. [30] I. H. Witten and E. Frank, E., Data Mining: Practical machine learning tools with Java implementations, San Francisco: Morgan Kaufmann, 2000. [31] D. Wolpert, “Stacked generalisation,” Neural Networks, 5, pp. 241-259, 1992.

9907

Eager and Lazy Learning Methods in the Context of ...

precipitation, runoff, water stages, etc. are of major focus in hydrologic modeling. Apart from the methods based on the detailed description of the physical processes (process or physically-based), the methods based on computational intelligence (CI) (often referred to as data-driven models) are gaining popularity [26].

316KB Sizes 1 Downloads 122 Views

Recommend Documents

Eager and Lazy Learning Methods in the Context of ...
at http://www.ics.uci.edu/~mlearn/MLRepository.html. [5] M. Bray and D. Han, ... AI'92, 5th Australian Joint Conference on Artificial. Intelligence, Edited by Adams ...

SPROUT: Lazy vs. Eager Query Plans for Tuple ...
query Q and a probabilistic database D, compute the distinct possible tuples in the ... the probability of 1OF formulas by mapping AND into product and OR into ...

Lazy Learners at work: the Lazy Learning Toolbox ... - ULB CoDE
In Section 2 we will introduce the Lazy Learning method in formal ... O nce obtained the local first-degree polynomial approximation, a prediction ... O n the basis of this information a final prediction .... collaboration of Iridia with the H oneywe

SPROUT: Lazy vs. Eager Query Plans for Tuple ...
2Department of Computer Science, Cornell University, Ithaca, NY 14853, USA. Abstract—A paramount challenge in probabilistic databases is the scalable ...

The Usability of Ambiguity Detection Methods for Context-Free ...
Problem: Context-free grammars can be ambiguous ... Overview. 1. Ambiguity in Context-Free Grammars. 2. .... Architecture and Software Technology, 2001.

Herodotus-In-Context-Ethnography-Science-And-The-Art-Of ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Herodotus-In-Context-Ethnography-Science-And-The-Art-Of-Persuasion.pdf. Herodotus-In-Context-Ethnography-Sci

The Lazy Editor.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. The Lazy Editor.

Comparing Machine Learning Methods in Estimation of Model ...
Abstract− The paper presents a generalization of the framework for assessment of predictive models uncertainty using machine learning techniques. Historical ...

in context academia in context programming - Research at Google
Think Programming,. Not Statistics. R is a programming language designed to work with data ... language—human or computer—can be learned by taking a ... degree of graphics but ones that are not ... Workshops and online courses in R are.

Multiculturalism and education for citizenship in a context of ...
Multiculturalism and education for citizenship in a context of neoliberalism.pdf. Multiculturalism and education for citizenship in a context of neoliberalism.pdf.

UNSUPERVISED CONTEXT LEARNING FOR ... - Research at Google
grams. If an n-gram doesn't appear very often in the training ... for training effective biasing models using far less data than ..... We also described how to auto-.

Learning Context Free Grammar Rules from a Set of Programs
We propose a technique which infers grammar rules from a given set of programs and an approx- ...... Semi-automatic Grammar Recovery. Software—. Practice ...

AFFECT AND RISK PERCEPTION IN THE CONTEXT ...
researchers [5], Canadian participants were presented with a scenario of a nuclear blast in ... to the statistical risk yet highly representative of the iconic images.

The Psychosocial Context of Trauma in Treating PTSD Patients.pdf ...
of posttraumatic stress disorder for decades (3). Since in- somnia has been observed in 90% of PTSD cases (4) and. nightmares related to the trauma in 70% (with or without. comorbid depressive disorder [5]), this is understandable. Both pharmacologic

Inferential statistics in the context of empirical cognitive ...
to decide how to deal with that variability, and how to characterize in some .... the ones you see in the business or weather report pages of your newspaper), and averages ..... compare the performance of a group of participants who took drug G with