XXXI IAHR CONGRESS 715

QUANTIFYING UNCERTAINTY OF FLOOD FORECASTING USING DATA DRIVEN MODELS DURGA LAL SHRESTHA1 and DIMITRI P. SOLOMATINE2 1

Research Fellow Associate Professor UNESCO-IHE Institute for Water Education P.O. Box 3015, 2601 DA Delft, The Netherlands 1 ( Tel: +31-15-2151-764, Fax: +31-15-2122-921, e-mail: [email protected]) 2 ( Tel: +31-15-2151-715, Fax: +31-15-2122-921, e-mail: [email protected]) 2

Abstract Flooding is a complex and inherently uncertain phenomenon. Consequently forecasts of it are inherently uncertain in nature due to various sources of uncertainty including model uncertainty, input uncertainty and parameter uncertainty. Several approaches have been reported to quantify and propagate uncertainty through flood forecasting models using probabilistic and fuzzy set theory based methods. In this paper, a method for quantifying uncertainty in flood forecasting using data driven modeling techniques is presented. Uncertainty of the model outputs is estimated in terms of prediction intervals. First, prediction intervals for training (calibration) data using clustering techniques to identify the distinguishable regions in input space with similar distributions of model errors were constructed. Secondly, a data driven model was built using the computed prediction intervals as targets. Thirdly, the constructed model was validated by estimating the prediction intervals for unseen (test) data. The method was tested on Sieve and Bagmati river basins using various data driven models. Preliminary results show that the method is superior to another method (uniform interval method) estimating prediction intervals. Keywords: Flood forecasting; Uncertainty; Prediction intervals; Data driven models; Models trees; Neural networks; Lumped conceptual rainfall-runoff model 1. INTRODUCTION Since flood is a complex and uncertain phenomenon, forecast of it is also inherently uncertain in nature. Incorporating the uncertainty estimation in the forecast can help the decision maker within the flood forecasting and warning system and thus enhance the reliability and credibility of both the forecasting and the warning system in real time flood management. Quantifying uncertainty within the flood forecasting enables an authority to set risk-based criteria for flood warning, provides information necessary for making rational decisions and offers potential for additional economic benefits of forecasts to every rational decision maker (Krzysztofowicz, 2001).

716 September 11~16, 2005, Seoul, Korea The uncertainties in flood forecasting have four important sources (e.g. Refsgaard and Storm, 1996): a) input data (e.g. precipitation, temperature, etc.) b) data used for calibration (e.g. runoff) c) model parameters d) imperfection of model structures. The error sources a) and b) depend on the quality of data, whereas c) and d) are more model specific. Uncertainty in flood forecasting can be estimated using several approaches. The first approach is to forecast the model outputs probabilistically (Krzysztofowicz, 2001) which is based on Bayesian framework. The second approach is to estimate uncertainty by analyzing the statistical properties of the model errors that occurred in reproducing the observed historical data. This approach has been widely used in statistical (Harnett and Murphy, 1980; Wonnacott and Wonnacott, 1996) and machine learning communities (Nix and Weigend, 1994; Heskes, 1997) for time series forecasting. In this approach uncertainty is estimated in terms of confidence intervals or prediction intervals. Recently Montanari and Brath (2004) proposed meta-Gaussian (Kelly and Krzysztofowicz, 1997) approach for the estimation of rainfall-runoff simulation uncertainty. It estimates the probability distribution of the model error conditioned by the value of the contemporary simulated river flows. The third approach is to use simulation and re-sampling based techniques such as Monte Carlo (Kuczera and Parent, 1998) and Generalised Likelihood Uncertainty Estimation (GLUE) (Beven and Binely, 1992) methods. The fourth approach is based on fuzzy theory based method (Maskey et al., 2004). This approach provides a non-probabilistic approach for modelling the kind of uncertainty associated with vagueness and imprecision. The first and the third approaches mentioned require the prior distributions of the uncertain input parameters or data to be propagated through the model to the outputs. The second approach requires certain assumptions about the data and the errors, and, obviously, the relevancy and accuracy of such approach depends on the validity of these assumptions. The last approach requires knowledge of the membership function of the quantity subject to the uncertainty. In this paper, we propose a new method for quantifying uncertainty in flood forecasting. Ideologically, it is close to the second approach i.e. uses the historical model residuals. Since the model residuals result from various sources of uncertainties as mentioned above, the method aims at estimating their aggregate effect without attempting to separate each individual contribution. In the situations when distributions of parameters or distributions of model errors are not known, we propose to use data driven modelling techniques to quantify uncertainty of flood forecasting in the form of prediction intervals (PIs). Data driven modelling induce a causal relationship between sets of input and output data in the form of a mathematical device, which in general is not related to the physics of the real world situation (Solomatine and Price, 2004). Various data driven techniques, such as linear regression, locally weighted regression (LWR), M5 model trees (MT) (Quinlan, 1992) and neural networks (ANN) can be used to predict the PIs using all or part of input vectors that are used to train (calibrate) the forecasting models. Upper and lower PIs are computed separately on the basis of model errors observed during training (calibration) of the forecasting models. Fuzzy C-means clustering (Bezdek, 1981) technique is used to determine zones of the input

XXXI IAHR CONGRESS 717 space having similar distributions of model errors. The PIs of each cluster are determined on the basis of empirical distributions of the errors associated with instances belonging to this cluster. The PIs of each instance are then determined according to grade of their membership in each cluster. After training of the model that is used to predict the PIs, they are applied to estimate the PIs for the forecasts made by the four data driven models (linear regression, LWR, MT, and ANN) in Sieve (Italy) river basin and Simulation made by the two-tank lumped conceptual rainfall-runoff model (Sugawara, 1995) in Bagmati (Nepal) river basin (DHM, 2000). Their outputs are compared to uniform interval method (UIM) as benchmark method using prediction intervals coverage probability (PICP) and mean prediction interval (MPI). The presented method does not need any information about the distribution of parameters and is not based on any assumptions about the distribution of errors. Most of the existing methods estimate upper and lower PIs, which are symmetrical about the point estimate. The presented method, however, computes upper and lower PIs independently and they are not necessarily symmetrical. The method is robust and flexible and it can be applied to the outputs of any models regardless of its class or structure. 2. PREDICTION INTERVALS Prediction intervals (PIs) are intervals within which a future unknown value (e.g. point forecast) is expected to lie with a prescribed probability. The upper and lower limits are called prediction limits or bounds. The prescribed probability is called confidence level. PIs should be distinguished from the Confidence Interval (CI). The CI deals with the accuracy of our estimate of the true regression (statisticians formulate this differently - as estimation of the mean value of the outputs). In contrast, the PI deals with the accuracy of our estimate with respect to the observed value. It should be noted that the PI is wider than the CI. The PIs are widely used and present a comprehensible form of uncertainty measure. The PIs can be easily derived from probability density function and even from variance of the quantity subject to uncertainty. Most of the methods to construct a 100(1- α )% prediction limits for model output y used in practice are essentially of the following form: PLU = y + z α / 2 σ e (1) PLL = y − z α / 2 σ e where PLU and PL L are the upper and lower prediction limits respectively, z α / 2 is the value of the standard normal variate with cumulative probability level of α / 2 , and σ e is the standard deviation of the errors. Computation of σ e2 is not so easy in most of practical situation especially for multivariate models containing many equations or depending on nonlinear relationships. Nevertheless, bootstrap (Efron, and Tibshirani, 1993) techniques have been used in estimating uncertainty of neural networks models (e.g. Nix and Weigend, 1994; Heskes, 1997) to estimate σ e2 . These techniques are generally computationally time consuming.

718 September 11~16, 2005, Seoul, Korea 3. METHODOLOGY Due to the various sources of uncertainty as mentioned in the section 1, it is not surprising that model outputs do not match the observed values well. The essence of the proposed method is that historical residuals between the model outputs and the corresponding observed data are the best available quantitative indicators of the discrepancy between the real-world process and its model. These residuals between the model outputs and observed values give the valuable information that can be used to assess the model uncertainty. These residuals are often function of the model input’s values and can be modelled. The method identifies n zones (regions or clusters) in input space reproducing different distributions of historical residuals. It is assumed that a region in input space that is associated with any particular cluster has similar residuals or residuals with similar distributions. Having identified the clusters, the PIs for the particular cluster are computed from empirical distributions of the historical residuals that belong to the cluster under consideration. For instance, in order to construct 100(1- α )% PIs, α / 2 *100 and (1 − α / 2) *100 percentile values are taken from empirical distributions of residuals for lower and upper PIs respectively. Typical value for α is 0.05, so this corresponds to 95% interval. If input space is divided into crisp clusters, e.g., by K-means clustering method, and each instance belongs to only one cluster, this computation is straightforward. However, in case of fuzzy clustering e.g., by Cmeans method where each instance belongs to more than one cluster and is associated with membership grades that indicates the degree with which the instance belong to a given cluster, the computation of the above percentiles should take this into account. The following expression gives the lower PI, denoted by PIC L for cluster j: PIC L ( j ) = e(i ) i

N

k =1

k =1

(2)

i : ∑ µ j (k ) < α / 2 ∑ µ j (k )

where i is the maximum value of it that satisfies the above inequality, e(i) is the error (residuals) associated with the instance i (instances are sorted in ascending order) , µ j (i ) is

the membership grade of ith instance to cluster j. Similar type of expression can be obtained for the upper PI ( PICU ). This is illustrated in Fig. 1. ∑ µi

N

∑ µi i =1

N

(1 − α / 2) ∑ µ i i =1

N

α / 2 ∑ µi i =1

-2.65 -1.56 -1.45 -1.45 -0.23

0.58 1.12

Lower interval

1.13 1.85

2.01 2.69

e

Upper interval

Fig. 1 Computation of prediction intervals in case of using fuzzy C-means clustering.

XXXI IAHR CONGRESS 719 Once the PIs are computed for each cluster, the PIs for each instance in input space can be computed. Computation of the PIs for each instance also depends upon the clustering techniques employed. For example, if K-means clustering is employed, then the PI for each instance in the particular cluster is the same as that of the cluster. However, if fuzzy C- means clustering is employed, PIs are computed using the weighted mean of PIs of each cluster as: c

PI L (i ) = ∑ µ j (i ) PIC L ( j )

(3)

j =1

where PI L (i ) is the lower PI for ith instance, c is number of clusters. Similar type of expression can be obtained for the upper PI. Once PIs for each input instance is computed, it is possible to construct mapping function f PI that estimate underlying functional relationship between input x and target or desired PIs

as: PI = f PI (x)

(4)

where PI without subscripts denotes the PI in general. Mapping function f may be of any form from linear regression to non-linear functions such as ANN. In other words, given a set of N data pairs {x(n), PI (n)} , n = 1,…, N, we can train data driven model such as ANN to estimate underlying function relating x to PI . It is to be noted that the target or desired variable might be either intervals or limits. Mapping function f PI will be referred to as the prediction intervals model (PIM) in this paper. The performance of the PIM is evaluated using the prediction interval coverage probability (PICP). Even though target values of the PIs are not available in the validation data set, it is interesting to know whether the observed values (e.g. runoff) from the forecasting model are inside the estimated prediction bounds. By definition, the prediction bounds enclose the true but unknown value ( 1 − α )% of times on average (typically 95%). The PICP is the probability that the target of an input vector lies within the computed prediction bounds and is estimated by the corresponding frequency as follows: count (i ) V i : PL L ≤ t (i ) ≤ PLU PICP =

(5)

where V is number of data in the validation se, t is target value for which prediction bounds are constructed. If the clustering techniques and the PIM are optimal, then the PICP value will be consistently close to the ( 1 − α )%. Yet another performance measure for PIs can be introduced here as well. This is the mean prediction interval (MPI), which is an average prediction interval on validation data set and it measures the ability to enclose target values inside prediction bounds on cost of its width. MPI can be estimated by 1 V MPI = ∑ [ PI L (i ) + PI U (i )] (6) V i =1

720 September 11~16, 2005, Seoul, Korea 4. EXPERIMENTS AND RESULTS 4.1 STUDY AREA The method was applied in the two-study areas: Sieve river basin and Bagmati river basin. The Sieve river basin is a tributary of the Arno river basin having length of 56 km and located in the Tuscany region of Italy. The Arno basin that includes the Sieve catchment was extensively studied and used as a case study for various physically-based hydrologic modelling exercises (e.g. Todini, 1996). Data-driven methods were used by Solomatine and Dulal (2003). It has a drainage area of 822 km2. The basin covers mostly hills and mountainous areas. The climate of the basin is temperate and humid. Three months of hourly discharge (Q) and precipitation (R) data were available (Dec. 1959 to Feb. 1960, 2160 data points) which represent various types of hydrological conditions. Bagmati river basin is a medium sized river basin of Nepal with a catchment area of about 3700 km2 up to the Nepal-India border. It originates from the southern slope of Shivapuri lek (Mahabharat within Kathmandu valley) passes through the inner Mahabharat range and stretches to the plains of Terai (ending at Nepal-India border). The available data consists of 3 rainfall stations, 1 discharge station and 1 climatic station for air temperature. The daily time series data for rainfall, air temperature and discharge for 8 years (from 1988 to 1995, 2920 data points) were collected and analysed. The mean daily arial precipitation was calculated using a Thiessen polygon method and the daily evapotranspiration was computed using the modified Penman method recommended by FAO. 4.2 PROCEDURE Whole data set was split into training (calibration) and validation parts. Same training data set used in forecasting or simulation model was also used in training PIM. Similarly, same validation data set used in forecasting model was again used in validating PIM. Furthermore clustering was employed to the same training data to construct PIs. For Sieve river basin validation data constitutes first 300 points starting 01/12/59, 07:00 to 13/12/59,18:00 while remaining 1854 data points starting from 13/12/59, 19:00 to 28/02/60, 00:00 constitutes the training dataset. For Bagmati river basin, 1998 training data is from period 03/01/88 to 22/06/93, and 922 validation data - from period 23/06/93 to 31/12/95. For Sieve river basin, various data driven models from on simple linear regression to nonlinear regression such as ANN were fitted on training data set to forecast river flows Qt+i several hours ahead (i=1, 3 or 6). The input variables to these forecasting models are based on the previous values of flow (Qt − τq ) and previous values of effective rainfall (Re t − τr ) , where τq

being between 0 and 2 hours and τr being between 0 and 5 hours depending upon lead

time of forecasts. Selection of input variable to the forecasting models are based on the linear correlation coefficient and understanding of physical insight between the input variables and sought target or desired variable (see Table 1). Then the trained model was used to forecast on unseen data set i.e. the validation data set.

XXXI IAHR CONGRESS 721 For Bagmati river basin, 2 tank rainfall-runoff conceptual model is calibrated using training data. The model has a total of 8 parameters, which were estimated by automatically optimizing using GLOBE (Solomatine, 2005). Having model outputs for the training and validation data sets, the PIM was constructed to estimate the PIs on validation data set as follows. Fuzzy C-means clustering technique was first employed to construct the PIs for each cluster in the training data set and then to construct the PIs for each instance in the training data set. Note that the input to the PIM may constitute all or only some of the variables used in the forecasting model. Table 1 shows the description of input variables to the forecasting or simulation model, clustering, and the PIM. The target variables to the PIM are the upper and the lower prediction bounds. Alternately, the target variable may be the upper and the lower PIs, but in this case, predicted values from the forecasting model have to be added to obtain the prediction bounds. The PIs were constructed for 95% confidence level unless specified. The performance of the PIM is assessed by the PICM and the MPI introduced in Section 3. The optimal number of clusters in Fuzzy C-means clustering was identified using the performance index P (Amiri, 2003) and the Xie-Beni index S (Xie, and Beni, 1991). Sometimes the minimal values of these indices are not so pronounced, so the experiments were repeated with different numbers of clusters. Table 1. Input data to the models Case study Sieve river basin

Bagmati river basin

Designation SieveQ1

Target Variables Qt+1

SieveQ3

Qt+3

SieveQ6

Qt+6

BagmatiQ

Qt

Forecasting/Simulation Clustering Prediction Intervals Model Model Ret, Ret-1, Ret-2, Ret-3, Ret-4, SAME AS SAME AS Ret-5, Qt, Qt-1, Qt-2 FORECASTI FORECASTING NG Ret, Ret-1, Ret-2, Ret-3, Qt, SAME AS SAME AS Qt-1 FORECASTI FORECASTING NG Ret, Qt Same as Same as forecasting forecasting Rt, Et Rt, Et, Qt Rt, Rt-1, Et, Et-1, Qt-1

4.3 RESULTS AND DISCUSSION Fig. 2 shows clustering of input examples in Sieve river basin for 1 hour ahead prediction. Cluster 1 (C1) contains input examples with very high runoff, whereas cluster 5 (C5) is associated with very low values of runoff. Fuzzy C-means clustering is able to identify clusters corresponding to various mechanism of runoff generation process such high flow, base flow etc.

722 September 11~16, 2005, Seoul, Korea 800 C1 700

C2

600

C3 C4 C5

) s m ( ff o n u R

500

3/

400 300 200 100 0

0

200

400

600

800 1000 Time (hr)

1200

1400

1600

1800

Fig. 2 Clustering of input examples in training data set in SieveQ1 using fuzzy C-means clustering The comparison of performance (in terms of the PICP and the MPI) for SieveQ1 using various data driven models shows that performance of MT is better than that of the other models. The PICP for MT is 96.67% and MPI is 15.25 m3/s. We also compared the results with uniform interval method (UIM). The UIM constructs single PI from empirical distributions of errors on the whole training data and applied uniformly to the validation data set. The PICP using the UIM for MT is 91.13% and MPI is 11.18 m3/s. Simple model like linear regression also contains PICP close to 95%, but with very wide prediction bound.

Fig. 3 Computed prediction bounds for SieveQ1 using MT model. Black line corresponds to predicted runoff, while grey shading represents 95% prediction bounds. Small circles represent observed value out of computed prediction bounds

XXXI IAHR CONGRESS 723 Fig. 3 shows the computed prediction bounds for SieveQ1 on validation data set using MT. It is to be noticed that only few examples are outside the computed prediction bounds with reasonable width of prediction bound. We also computed the PICP and the MPI for SieveQ3 and SieveQ6. Although PICP are very close to 95%, MPIs are very wide as compared to SieveQ1 which is not surprising. Fig. 4 shows the deviation of the PICPs from the desired confidence level for Sieve river basin using MT. The PIs were constructed for various confidence levels ranging from 10% to 99%. It is to be noticed from Fig. 4 that the PICPs are very close to desired confidence levels at values higher than 80%, and in practice the PIs are constructed around this value. Furthermore, it can be noted that the PIs are too narrow in most of cases as the PICPs are below the straight line. Such evidence was also reported by Chatfield (2000). In these cases the PIM underestimates uncertainty of the model outputs. 100 SieveQ1

PICP (%)

80

SieveQ3 SieveQ6

60

ideal

40 20 0 0

20

40 60 80 100(1-alpha)% confidence level

100

Fig. 4 The PICP for different values of confidence level using MT model in Sieve river basin Fig. 5 presents fan chart showing MPIs for Sieve river basin in different forecast lead times for different degree of confidence levels. From figure it is evident that the width of the PIs increases with the increase of the confidence level. Moreover it is also illustrated that the width of PIs increases as forecast lead time increases. Thus uncertainty of model forecast increases as lead time increases.

Fig. 5 A fan chart showing mean prediction intervals for flow prediction in Sieve river basin upto 6 hours ahead. The darkest strip covers 10% probability and the lightest coves 99%

724 September 11~16, 2005, Seoul, Korea Table 2 shows the cluster centers and computed PIs in Bagmati river basin. It is to be noticed that cluster C3 corresponds to very high values of rainfall, evapotranspiration and runoff, while C5 is - to with very low values of these Computed PIs are consistent with values of inputs as C3 has high value of PIs, while C5 has low value of PIs. This fact is also supported by Table 3, which shows very high degree of correlation of computed PIs with rainfall and runoff. Table 2. Investigation of cluster centre and prediction intervals in Bagmati river basin Cluster

Cluster centre 3

Lower prediction

Upper prediction

ID

Rt (mm/day)

Et (mm/day)

Qt (m /s)

interval

interval

C1

0.96

0.78

57.98

-94.92

97.46

C2

2.92

1.16

116.62

-175.95

196.71

C3

28.72

1.30

515.51

-489.59

401.78

C4

4.37

1.38

153.95

-220.57

202.74

C5

0.38

0.51

27.96

-52.47

46.74

Table 3. Correlation coefficient of computed prediction intervals of cluster with value of cluster centers Inputs Lower prediction interval Upper prediction interval

Rt

Et

Qt

-0.965

-0.736

-0.990

0.927

0.799

0.964

Using 5 numbers of clusters, experiments were repeated to compute PIs for output of simulation model using different data driven models as conducted in Sieve river basin. The results show that MT is superior as compared to other data driven models. Only 5.1% of the validation data fall outside the computed 95% PIs (1.4% and 3.7% below lower and above the upper PI respectively) with MPI of 332.77 m3/s. Furthermore, these points are quite uniformly distributed along the range of the observed runoff. The UIM gives PICP of 92.6% with MPI of 389.25. Due to lack of space, supporting figures are not presented here. 5. CONCLUSIONS This paper presents a novel method for quantifying uncertainty of flood forecasting models using data driven modelling techniques. Uncertainty of flood forecasting is estimated in terms of the PIs. Computed PIs explicitly take into account all sources of uncertainty of the model outputs without attempting to separate the contribution given by the different sources of uncertainty. The methodology is independent of the structure or class of the forecasting model as it requires only the model outputs. Unlike existing techniques the methodology does not require the knowledge of prior distribution of parameters and does not rely on any assumptions about the distribution of errors. Most of the known techniques compute upper

XXXI IAHR CONGRESS 725 and lower prediction intervals symmetrically. The presented methodology computes upper and lower intervals independently. The methodology has been applied in order to estimate the PIs of the outputs of flood forecasting models in two case studies: Sieve and Bagmati river basins. The result shows that the computed prediction bounds are satisfactory as the PICP are very close to desired degree of confidence level with reasonable width of the prediction bounds. We also compare our results with the UIM and the results demonstrate that the new method performs consistently better than the UIM. ACKNOWLEDGMENT The authors acknowledge the valuable input of Mr. Javier Rodriguez during his M.Sc. study at UNESCO-IHE Institute for Water Education in 2003.This work is partly supported by the EU project Integrated Flood Risk Analysis and Management Methodologies (FLOODsite), contract GOCE-CT-2004-505420. REFERENCES Amiri, M. (2003). Yashil's Fuzzy C-Means Clustering MATLAB Toolbox, Ver. 1.0. Available: http://yashil.20m.com/. Beven, K.J., and Binely, J. (1992). The future of distributed models: model calibration and uncertainty prediction. Hydrological Processes, 6, 279-298. Bezdek, J.C. (1981). Pattern Recognition with Fuzzy Objective Function Algorithms. Kluwer Academic Publishers Norwell, MA, USA. Chatfield, C. (2000). Time-Series Forecasting. Chapman & Hall/CRC. DHM. (2000). Hydrological records of Nepal, streamflow summary, Dep. of Hydrology and Meteorology, Nepal. Efron, B., and Tibshirani, R. (1993). An Introduction to the Bootstrap. Chapman and Hall, New York. Harnett, D.L., and Murphy, J.L. (1980). Introductory Statistical Analysis. Addison-Wesley publishing compay. Heskes, T. (1997). Practical confidence and prediction intervals. Advances in Neural Information Processing Systems 9, M. Mozer et al. (eds.), MIT Press, 176-182. Kelly, K.S., and Krzysztofowicz, R. (1997). A bivariate meta-Gaussian density for use in hydrology, Stochastic Hydrology and Hydraulics, 11, 17-31. Krzysztofowicz, R., (2001). The case for probabilistic forecasting in hydrology. Journal of Hydrology, 249, 2-9. Kuczera, G., and Parent E. (1998). Monte Carlo assessment of parameter uncertainty in conceptual catchment models: the Metropolis algorithm, Journal of Hydrology, 211, 69-85. Maskey, S., Guinot, V., and Price, R.K. (2004). Treatment of precipitation uncertainty in rainfall-runoff modelling, Advances in Water Resources, 27, 889-898. Nix, D., and Weigend, A. (1994). Estimating the mean and variance of the target probability distribution, Proc. of the IJCNN, IEEE, 55-60.

726 September 11~16, 2005, Seoul, Korea Refsgaard, J.C., and Storm, B. (1996). Construction, calibration and validation of hydrological models. In Distributed Hydrological Modelling, Abbott, M.B., and Refsgaard, J.C. (eds.), Kluwer Academic Publishers. Todini, E. (1996). The ARNO rainfall–runoff model. Journal of Hydrology, 175, 339–382. Quinlan, J.R. (1992). Learning with continuous classes. Proc. of the 5th Australian Joint Conf. on AI, World Scientific, Singapore. Solomatine, D.P. (2005). Global Optimization Tool. Available: www.data-machine.com. Solomatine, D.P., and Dulal, K.N. (2003). Model tree as an alternative to neural network in rainfall-runoff modelling. Hydrological Science Journal, 48(3), 399–411. Solomatine, D.P., and Price, R.K. (2004). Innovative Approached to Flood Forecasting Using Data Driven and Hybrid Modelling. In Proc. of the 6th Int. Conference on Hydroinformatics, Liong et al. (eds.), World Scientific Publishing Company, Singapore. Sugawara, M. (1995). Tank Model. In Computer Models of Watershed Hydrology, Singh, V. P. (ed.), Water Resources Publication. Wonnacott, T.H., and Wonnacott, R.J. (1996). Introductory Statistics, John Wiley & Sons, Inc. Xie, X.L., and Beni, G. (1991). A Validity Measure for fuzzy clustering. IEEE Trans. on Pattern Analysis and Machine Intelligence, 13(8), 841-847.

quantifying uncertainty of flood forecasting using data ...

Proc. of the 5th Australian Joint Conf. on AI, World Scientific, Singapore. Solomatine, D.P. (2005). Global Optimization Tool. Available: www.data-machine.com.

244KB Sizes 4 Downloads 198 Views

Recommend Documents

A GIS-Based Flood Forecasting System - Esri
Mar 17, 2016 - manage, analyze, and serve the massive amounts of data ... merged with data from other Flemish organizations for advanced impact analysis ...

An Approach of Electric Power Demand Forecasting using Data ...
Data-Mining Method: a case study of application of Data-Mining technique to ... long range demand under the influence of population, big plant, building, etc,.

Plio-QUMP: Quantifying Uncertainty in Model ...
The mid-Piacenzian Warm Period is an interval of Earth history when global ... to examine the potential long-term effects of global warming through the.

Quantifying Organismal Complexity using a ... - Semantic Scholar
Feb 14, 2007 - stomatitis virus, and to illustrate the consistency of our approach and its applicability. Conclusions/Significance. Because. Darwinian evolution ...

Quantifying the Uncertainty about the Fit of a New ...
the Bankard Fund of the University of Virginia and the Swiss National Science Foundation is gratefully acknowledged. .... equation proposed by Sbordone that implies some degree of elasticity for firm-specific ..... case, the model does a poor job exp

Uncertainty Reduction of Damage Growth Properties Using ... - UFL MAE
prognosis techniques, it is necessary to incorporate the measured data into a damage .... with initial half-crack size ai subjected to fatigue loading with constant ...

Quantifying Hardware Security Using Joint Information ...
Department of Computer Science and Engineering ... encryption” or “a hardware security module makes the system .... rate of flow between variables [27].

hierarchical forecasting of web server workload using ...
We propose a solution to the web server load prediction problem based on a ... web applications and services, and maintained by a host service provider.

Quantifying Organismal Complexity using a Population ...
Feb 14, 2007 - ... of California San Diego, La Jolla, California, United States of America, 3 Eidgenössische Technische ..... which we list here in parentheses.

Rate my data: quantifying the value of ecological data ...
2Woods Hole Research Center, 149 Woods Hole Road, Falmouth, Massachusetts 02540 USA ... data do we need?'' These questions are relevant not only for model developers, who need observational data to improve, constrain, and test their models, but also

Data integration with uncertainty
Feb 17, 2008 - This section describes the requirements from a data integra- tion system that supports uncertainty and the overall archi- tecture of the system.

Data integration with uncertainty - Springer Link
Nov 14, 2008 - sources by automatic methods (e.g., HTML pages, emails, blogs). ..... If a tuple is an answer to Q under multiple mappings in m, then we add.

Flood Routing in Ungauged Catchments Using ...
... and Environmental Hydrology, University of KwaZulu-Natal, Private Bag X01 Scottsville, 3209, Pietermaritzburg, South ... average velocity (m.s-1), and.

Quantifying kids
Children's problems with quantified sentences involve a number of factors. One important factor is the ...... bridge, Mass. Braine, M.D.S. and B. ... Annual Boston University Conference on Language Development, 184-195. Freeman, N.H., C.G. ...

Minister launches new flood data on ACTMAPi.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. 131217 - media ...

Data-driven modelling vs. machine learning in flood ...
Data-driven modelling vs. machine learning in flood forecasting. D.P. Solomatine, B. Bhattacharya and D.L. Shrestha. UNESCO-IHE Institute for Water Education. P.O. Box 3015, Delft, The Netherlands. ({d.solomatine, bhatt7, shres15}@unesco-ihe.org / Ph

Motion planning under bounded uncertainty using ...
Department of Electrical and Computer Engineering. University of Illinois at ... system, show that the position of this ensemble is controllable, and derive motion ..... apply the Lie algebra rank condition because the configuration space is no ...

Using HMM for Forecasting Stock Market Index
The Australian Stock Exchange Limited (ASX) was formed in 1987 by legislation of the Australian Parlia- ment [2]. The stock markets are now an integral part of.

Quantifying Residual Finiteness
Aug 3, 2015 - Then FL. H(n) ≼ FS. G(n). Proof. As any homomorphism of G to Q restricts to a ..... Mathematics Department, Queen Mary College, Lon-.

Long-term Forecasting using Tensor-Train RNNs
One of the central questions in science is forecasting: given the past history, how well can we predict the future? In many domains with complex multivariate correlation structures and nonlinear dynamics, forecasting is highly challenging since the s