Startseite A Comparison of Three Soft Computing Techniques, Bayesian Regression, Support Vector Regression, and Wavelet Regression, for Monthly Rainfall Forecast
Artikel Open Access

A Comparison of Three Soft Computing Techniques, Bayesian Regression, Support Vector Regression, and Wavelet Regression, for Monthly Rainfall Forecast

  • Ashutosh Sharma und Manish Kumar Goyal EMAIL logo
Veröffentlicht/Copyright: 17. September 2016
Veröffentlichen auch Sie bei De Gruyter Brill

Abstract

Rainfall, being one of the most important components of the hydrological cycle, plays an extremely important role in agriculture-based economies like India. This paper presents a comparison between three soft computing techniques, namely Bayesian regression (BR), support vector regression (SVR), and wavelet regression (WR), for monthly rainfall forecast in Assam, India. A WR model is a combination of discrete wavelet transform and linear regression. Monthly rainfall data for 102 years from 1901 to 2002 at 21 stations were used for this study. The performances of different models were evaluated based on the mean absolute error, root mean square error, correlation coefficient, and Nash-Sutcliffe efficiency coefficient. Based on model statistics, WR was found to be the most accurate followed by SVR and BR. The efficiencies for the BR, SVR, and WR models were found to be 32.8%, 52.9%, and 64.03%, respectively. From the spatial analysis of model performances, it was found that the models performed best for the upper Assam region followed by lower, southern, and middle regions, respectively.

1 Introduction

Rainfall is one of the most important components of the hydrological cycle. It has an importance in many fields, including agriculture, irrigation, hydroelectric production, scientific research, etc. Many catastrophic events like flood and drought are directly linked to rainfall and, hence, it becomes extremely important to have an accurate forecast of rainfall. Rainfall is highly stochastic in nature and is dependent on many local and global atmospheric processes, which makes rainfall prediction one of the most challenging problems around the globe.

Generally, two different approaches are used for rainfall prediction: empirical and dynamical. In the empirical approach (e.g. regression, artificial neural networks, fuzzy logic and other machine learning and data mining techniques), the past records of rainfall, atmospheric parameters, and oceanic parameters are used to capture the trend of rainfall w.r.t. to governing parameters. In the dynamical approach, physical models based on some system of equations are used to predict the rainfall and other atmospheric parameters based on some initial condition provided [27]. The empirical approach has been widely used by researchers for rainfall prediction [19, 22].

For this study, we have used three different regression techniques, viz. Bayesian regression (BR), support vector regression (SVR), and wavelet regression (WR), for rainfall forecasting. BR tries to model the uncertainty in unknown parameters by assigning a probability distribution, and then Bayes’ theorem is used to find the posterior distribution. It intends to find a value for unknown parameters that maximizes the posterior distribution. Many applications of Bayesian methods in different fields of science and engineering are available in the literature [4, 5, 8, 15, 23, 28, 32]. A Bayesian framework is used to analyze a daily rainfall series for estimation of extreme rainfall events, and it is shown that use of prior expert information can supplement data and can improve the estimate of extreme events [5]. A Bayesian network is used for modeling the spatial and temporal dependencies among the 100 stations, and the network is combined with the numerical atmospheric prediction to forecast the precipitation [4]. A multivariate relevance vector machine model can be used to forecast releases from a system of multiple reservoirs with good accuracy in learning the input-output patterns [32].

Support vector machine (SVM) was introduced by Vapnik [33] based on the statistical learning theory (SLT). It tries to find a decision rule that provides good generalization by selecting some data points in the training data called support vectors. SVM methods have been widely used in the field of hydrology [1, 2, 7, 20, 25, 30, 38, 39]. Hong and Ping-Feng [12] made an assessment of the support vector regression (SVR) techniques in rainfall prediction, and showed that SVR could be a better alternative for rainfall forecasting. A hybrid support vector technique was used to forecast two hydrological time series [30]. Wei [37] used a coupled wavelet SVMs for hourly rainfall forecast during the tropical cyclones.

Wavelet transform has been used for investigating the variation, periodicity, and trends in data time series in many studies. Discrete wavelet transform (DWT) can effectively be utilized to classify distinct hydroclimatic categories [31]. The wavelet methods are used to model rainfall and runoff for daily and half-hourly sampling rates of karstic springs, and are found able to discriminate the karstic rapid response and recharge, from slower infiltration responses [18]. Chou and Wang [3] applied DWT to decompose unit hydrograph and utilized updated wavelet coefficients of unit hydrograph to accurately predict one-step-ahead runoff. Dai et al. [6] used wavelet analysis to study the decay in summer monsoon in north China over the interdecadal time scale. Zhou et al. [40] have shown that a wavelet predictor-corrector model performed more accurately compared to the ARMA model, seasonal ARIMA model, and BP artificial neural network for the prediction of discharge. Kisi [16] compared the performance of the WR model to the artificial neural network (ANN) and autoregressive (AR) models for streamflow prediction, and found that WR outperformed the ANN and AR models. The WR model was found superior to the ANN model for prediction of daily river stage [17]. A comparison of the WR model and ANN has shown that the WR model provided better monthly rainfall forecast [9]. Many studies have shown SVR models to be better than ANN models [1, 7, 20, 29]. The WR model decomposes the monthly rainfall time series into detail and approximation components using DWT, and a new time series is generated by adding an approximation component and effective detail components. This new time series is given as input to the linear regression (LR) model, keeping the original rainfall time series as output. Some hydrological applications of WR have revealed that WR can perform better than non-linear models like ANN, etc., in terms of accuracy [16, 17]. In the present research work, the accuracy of WR in rainfall prediction is compared with that of SVR and BR.

Assam is a northeastern state of India, where agriculture is the backbone of state economy. The northeastern region of India is one of the most rainfall-receiving regions in the world. Flood occurs in many parts of the state every year, causing huge loss in terms of infrastructure and human lives. An advance prediction of extreme rainfall events causing floods can help in reducing the impact of floods. The aforementioned three soft computing techniques are used in this paper for rainfall forecast at 21 stations in Assam, India. The BR approach finds the posterior distribution of regression coefficients and averages over a large number of samples. The Bayesian averaging is effective in avoiding the overfitting of the training data. Vapnik’s ε-insensitive loss function is used for the SVR, and the Gaussian radial basis function is used for non-linear mapping of input data space. A coupled WR and LR approach is used for rainfall forecast. The present study compares six different models developed based on the aforementioned three soft computing techniques.

2 Study area and data

2.1 Study Area

Assam is a northeastern state of India that is located between latitudes 24°8′N to 28°2′N and longitudes 89°42′E to 96°E. The location of the state on the India map is shown in Figure 1. The state is located at the foothills of the eastern Himalayas and receives an average annual rainfall of around 300 cm from the southwest monsoon. There are around 51 forests and subforests in the states because of its topographical and climatic features. The state experiences heavy rainfall and floods every year, which cause riverbank erosions and landslides in many parts of the state. Rainfall and floods also affect agriculture, which accounts for more than a third of the state domestic product and is the principal occupation of the rural people, who constitute nearly 90% of the total population [41].

Figure 1: Location of Assam and Stations on India Map, and Distribution of Monthly Mean Rainfall over Assam.The numbers represent the stations.
Figure 1:

Location of Assam and Stations on India Map, and Distribution of Monthly Mean Rainfall over Assam.

The numbers represent the stations.

There is significant spatial and temporal variation in rainfall over the state due to topographic and geographic features. Southwest monsoon brings heavy rainfall during the June to August period and causes flood in many parts of the state. The state suffers huge monetary losses every year because of heavy rainfall and floods. The amount of rainfall received by some regions in monsoon months is >800 mm. During the non-monsoon period, the monthly mean rainfall remains much smaller. A box plot showing the annual mean rainfall received by 21 stations in Assam between 1901 and 2002 for every month is presented in Figure 2. Due to this high temporal variability, prediction of rainfall becomes very difficult.

Figure 2: Box Plot Showing Annual Mean Rainfall at All Stations for Every Month Between 1901 and 2002.
Figure 2:

Box Plot Showing Annual Mean Rainfall at All Stations for Every Month Between 1901 and 2002.

2.2 Data

The monthly mean precipitation data of 21 stations spread over the whole state is used for this study. Data for precipitation were obtained from the India water portal website (http://www.indiawaterportal.org/met_data/). The monthly mean rainfall is available for 102 years from January 1901 to December 2002. The distribution of rainfall is shown in Figure 1. The eastern part of the state receives less rainfall, compared to other parts of the state. The rainfall statistics of 21 stations for 102 years data are presented in Table 1. The maximum recorded monthly rainfall is 1594.27 mm at Kamrup station, which occurred in July 1974. The mean monthly rainfall varied between 172 and 294 mm. The standard deviation varied between 155.35 mm at station no. 21 (Tinsukia) and 301.43 mm at station no. 11 (Kamrup). The skewness varied between 0.67 at Lakhimpur station and 0.92 at Hailakandi station. Skewness is positive at all the stations, which shows that the frequency distribution of rainfall is asymmetric and it lies on the right side of the mean. The distribution of the data is less scattered as the values of skewness become lower. Out of 102 years data, the first 60 years data (from January 1901 to December 1960) were used for calibrating the models and the remaining 42 years data (from January 1961 to December 2002) were used for validation.

Table 1:

Geographical Latitude, Longitude, and Rainfall Statistics of 21 Stations.

Station no.Station nameLatitudeLongitudeMonthly mean rainfall
MinimumMaximumMeanStandard deviationSkewness
1Jorhat26°45′N94°13′E0.19677.06179.26162.860.7
2Barpeta26°19′N91°00′E01033.69230.73239.380.83
3Cachar25°05′N92°55′E01293.43275.71265.960.87
4Darrang26°45′N92°30′E01069.2234.79237.250.83
5Dhemaji27°29′N94°35′E0.2685.51180.44160.980.69
6Dhubri26°02′N89°58′E01020.58185.85198.970.88
7Dibrugarh27°29′N94°54′E0.2703.83183.69164.320.73
8Goalpara26°10′N90°37′E01108.47222.13231.720.83
9Golaghat26°31′N93°58′E0.06705.11186.69170.580.69
10Hailakandi24°41′N92°34′E01323.74261.82256.930.92
11Kamrup26°11′N91°44′E01594.27292.04301.430.89
12Karbi Anglong26°00′N93°30′E0.031234.26270.23265.140.87
13Karimganj24°52′N92°21′E01197.53233.98230.070.91
14Kokrajhar26°24′N90°16′E0997.85208.74221.320.83
15Lakhimpur27°14′N94°6′E0.13674.85178.1162.080.67
16Nagaon26°21′N92°41′E01525.96293.95294.850.89
17Nalbari26°25′N91°26′E01055.19231.85239.490.85
18North Cachar Hills25.18°N93.03°E01233.48294.86280.440.8
19Sibsagar26°59′N94°38′E0.19694.85179.75161.60.71
20Sonitpur26°38′N92°48′E0784.22192.8185.470.71
21Tinsukia27°30′N95°22′E0.19722.23172.87155.350.78

3 Methodology

A general LR problem can be described in the form of a target vector (t), a set of input vectors (xn), and a functional relationship (f) between t and x, which also includes some additive noise components. This can be represented as

(1)tn=f(xn,β)+ϵn,

where ϵ is an additive noise process in which ϵn are i.i.d. (independent identical distributions) and β is the vector of adjustable parameters. The main objective of all regression techniques is to set a selection criterion for the function f such that it fits the maximum data points in the training data set and gives a good generalization for new data points. The widely used least square technique is based on error function minimization, in which it tries to minimize the sum-of-squares error and set β* for the least sum-of-square errors. Error function minimization techniques suffer from overfitting of training data and poor generalization. Limiting the model complexity can help in avoiding the overfitting; however, the model can still suffer from poor generalization in case it is not flexible enough to capture the trend of the data. In the maximum likelihood technique, a normal distribution with zero mean and variance σ2 is assumed for the additive terms (ϵn) in Eq. (1) so that p(ϵ|σ2)=N(ϵ|0, σ2) and we select the value of β*, which maximizes the likelihood function.

3.1 Bayesian LR

In BR, the uncertainty in β is represented by a probability distribution, p(β). Bayes’ theorem is used to express the posterior distribution for β as the product of prior distribution and likelihood function:

(2)p(β|t,α,σ2)p(β|α)p(t|β,σ2).

The value of β is chosen, which maximizes the posterior distribution. We integrate over the distribution of β to make predictions. The integration over the distribution overcomes the problem of overfitting, as we are averaging over a large number of possible solutions. The maximization of the posterior distribution in Eq. (2) is equivalent to the minimization of the negative logarithm of the distribution. Different samples of the solution are generated using the mean and variance of the distribution. Solving the minimization problem, the posterior can be expressed in terms of mean and variance.

For this study, we have assumed a Gaussian prior on β (regression coefficients) with mean zero and covariance matrix Λ−1 so that β~N(0, Λ−1), where Λ is the inverse covariance or precision matrix that is used to make computation easier. The posterior distribution on β can be represented as

(3)β=N(μβ,Σβ),

where

Posterior mean, μβ=(XTX+σ2Λ)−1XTy,

Covariance matrix, Σβ=σ2(XTX+σ2Λ)−1.

This model works in two steps. In the first step, the mean and covariance matrix for posterior distribution of β is calculated using Eq. (3). In the second step, β is approximated from this normal distribution by averaging over a large number of samples generated by using multivariate normal distribution:

(4)β~MVN(μβ,Σβ).

3.2 Support Vector Regression

SVM is a method of supervised learning that is based on SLT. It was developed by Vapnik and his team, and is generally used for classification and regression analysis [33, 34]. The most important advantage of SVM is the principle of structural risk minimization, which is shown to be better than traditional empirical risk minimization [11, 24]. It tries to minimize the upper-bound generalization errors instead of minimizing the local training errors. It tries to find a decision rule that provides better generalization by selecting some subset of training data called support vectors. To take into account the non-linearity present in the data, the kernel functions are used for non-linearly mapping the input space to a higher-dimensional feature space.

In SVR, a loss function (Lε(t, f(x))) is used, which describes the deviation between the function f(xi) outcomes and the actual target values ti, and allows some deviation between the target values ti and the function values f(xi). Vapnik’s ε-insensitive loss function is used, which has some algorithmic advantages like providing sparse decomposition and rendering the computation more amenable [35]. Figure 3 shows Vapnik’s ε-insensitive loss function. It is defined as

Figure 3: Vapnik’s ε-Insensitive Loss Function.
Figure 3:

Vapnik’s ε-Insensitive Loss Function.

(5)Lε(t,f(x))={0if|tf(x)|ε|tf(x)|εotherwise.

The points that lie outside the ε-insensitive zone (outside red dotted lines) add to the cost. The deviations outside the ε-insensitive zone are penalized in a linear manner. The associated convex quadratic programming problem is

(6)Minimize12||w||2+Ci=1n(ξi+ξi).
(7)Subject toti(w.xi+b)ε+ξi(w.xi+b)yiε+ξiξi0ξi0i=1,2,,n,

where ξi and ξ_i are slack variables. These variables represent the deviations for the training data points lying outside the ε-insensitive zone. The slack variables have zero value for all the data points lying inside the ε-insensitive zone and increase progressively for the data points lying outside the zone. C is the regularization constant that assigns a penalty when a training error occurs. The current optimization problem can be easily optimized in the dual form.

A dual set of variables, αi and α¯i, is introduced to build the Lagrange function from the primal objective function [Eq. (6)] and the corresponding constraints [Eq. (7)]. The dual formation of the optimization problem is

(8)Maximize12i,j=1nα_iα¯i(α_jα¯j)(xi.xj)εi=1n(α_i+α¯i)+i=1nyi(α_iα¯i)

Subject toi=1n(α_iα¯i)=00α_iC,i=1,2,,n0α¯iC,i=1,2,,n.

After determining αi and α¯i, the Karush-Kuhn-Tucker method is utilized to find the parameters w and b. Prediction in an LR function is expressed as

(9)f(x)=i=1nαiα¯ixi.xj+b.

To make SVR deal with non-linear cases, a function ϕ is used to map the decision space variable (X) to a higher-dimensional space. The dual problem, after applying this transformation, becomes

(10)Maximize12i,j=1nα_iα¯i(α_iα¯j)ϕ(xi).ϕ(xj)εi=1n(α_i+α¯i)+i=1nyi(α_iαi).
(11)Subject toi=1n(α_iα¯i)=00α_iC,i=1,2,,n0α¯iC,i=1,2,,n,

where k(xi, xj)=(ϕ(xi).ϕ(xj)) is called the kernel function. Any function can be used as kernel function if it satisfies Mercer’s theorem [34]. Gaussian radial basis function (GRBF) is used for this study, which can be defined as K(xi, xj)=exp(−γ||xixj||2).

The hyperparameters such as the SVR constant ε, regularization constant C, and σ (for radial basis function kernels) influence the performance of SVR. Parameter C allows a trade-off between training error and model complexity. If a model is not sufficiently complex, then it may fail to capture the underlying trend of the data and hence cause underfitting. On the other hand, if the model is too complex, then it may capture the noise of the data and, hence, can suffer from overfitting.

The magnitude of training error increases for the small values of C and hence tends to underfit the training data. However, a higher value of C will tend to a behavior as that of a hard-margin SVM and hence overfits the training data [13]. The parameter ε influences the number of support vectors by deciding the thickness of the ε-insensitive zone. Hence, its value affects both generalization capability and the complexity of the approximation function. A lower value of ε leads to higher number of support vectors and increases the complexity. A higher value of ε leads to smaller number of support vectors and results in more flat estimates of the regression function. The performance of SVR is sensitive to these parameters; hence, it is important to find appropriate values for C and ε [14]. Determining the appropriate value of these parameters is often a heuristic trial-and-error process [25]. It is shown that an exponentially growing sequence of parameters works better [29].

3.3 Wavelet Regression

Wavelet analysis is an advanced signal processing technique that has been extensively used in communications, data compression, acoustics, signal processing, image processing, etc. It was introduced by Grossmann and Morlet [10] in the early 1980s. The signal or function is expressed as the combination of “little waves” called wavelets. Unlike Fourier transform, where the signal is divided into smooth sinusoids of unlimited duration, the wavelet transform splits the signal into wavelets of finite duration and zero mean [16]. Wavelets are localized both in frequency and time domains. This property of wavelets is called time-frequency localization. Wavelet analysis features the analysis of temporal variations at different time scales, and can work on non-stationary signals as well. Fourier transform assumes infinite length signals and concentrates mainly on their frequencies only while wavelet transform can be applied to any kind and any size of the signal.

The wavelet function ψ(t) is defined as +ψ(t)dt=0. The term ψa,b(t) is obtained by compression and expansion of ψ(t):

(12)ψa,b(t)=|a|12ψ(tba)bR,aR,a0,

where R represents real numbers, ψa,b(t) is the successive wavelet, and a and b are the scale of frequency factor and time factor. When the term ψa,b(t) satisfies Eq. (12), for a finite energy signal [16, 26, 40], the successive wavelet transform of f(t) is

(13)Wψf(a,b)=|a|12Rf(t)ψ¯(tba)dt,

where ψ(t)¯ represents the complex conjugate function of ψ(t). f(t) is decomposed at a different resolution level (scale) in Eq. (13). In real applications, the successive wavelets are generally discrete. Let a=a0j,b=kb0a0j,a0>1, b0R, and k, j are integers. The DWT of function f(t) is defined as

(14)Wψf(j,k)=a0j/2Rf(t)ψ¯(a0jtkb0)dt.

The widely used values for parameter a0 is 2 and for b0 is 1 time step. When a0 is 2 and b0 is 1, Eq. (14) changes into binary wavelet transform [16] as

(15)Wψf(j,k)=2j/2Rf(t)ψ¯(2jtk)dt.

Wψf(a, b) or Wψf(j, k) reflects the frequency characteristics (a or j) and time characteristics (b or k) of the original time series. The parameter a decreases when the temporal resolution is high and the frequency resolution of the wavelet transform is low. On the other hand, a increases when the temporal resolution is low and the frequency resolution is high [36]. For discrete time series f(t) for different time steps (integral), the DWT is defined as [42]

(16)Wψf(j,k)=2j/2t=0N1f(t)ψ¯(2jtk),

where Wψf(j, k) is the wavelet coefficient for the discrete wavelet of scale a=2j and b=2jk.

The input signal (monthly rainfall time series) is decomposed into a certain number of sub-time series components (Ds) using the Mallat DWT algorithm [21]. Every component has an important role in the original time series, and all have distinct behavior. Thus, the contribution of different subcomponents to the original time series varies from each other [36]. The decomposition process consists of a number of successive filtering steps. The original signal is first decomposed into approximation and detail components. It is broken down into many lower-resolution components. The details are low-scale high-frequency components of the signal, while the approximations are the high-scale, low-frequency components (Figure 4). The higher scales consist of the extended version of a wavelet, and the coefficients refer to the slowly-changing coarse features of low-frequency components. The lower scales present the condensed wavelet and follow the rapidly changing details (high-frequency components) of the signal [21].

Figure 4: Signal Decomposition by DWT for Three Resolution Levels.St denotes the original signal, Ai denotes the approximation components, and Di denotes the detail components.
Figure 4:

Signal Decomposition by DWT for Three Resolution Levels.

St denotes the original signal, Ai denotes the approximation components, and Di denotes the detail components.

The WR model used in the present study is a combination of DWT and LR. DWT is utilized for preprocessing of the original monthly rainfall series before use in the LR model. The decomposed subcomponents (detail and approximation) are utilized as inputs for LR, keeping the original rainfall series as input. The rainfall time series is processed using DWT separately for both calibration and validation period.

3.4 Models

Using these three techniques and two different sets of input variables, six different models were developed. The details of these models are provided in Table 2. All the models were calibrated and validated for the same data, as specified in Section 2.2.

Table 2:

Different Models and Input Variables.

ModelApproach usedInput variables
WR1Wavelet regression
SVR1Support vector regressionR(t)=f(R[t−1])
BR1Bayesian regression
WR2Wavelet regression
SVR2Support vector regressionR(t)=f(R[t−1], R[t−2])
BR2Bayesian regression

4 Results and Discussion

Monthly rainfall at 21 stations in Assam is forecasted using three different techniques, namely BR, SVR, and WR. Out of 102 years of available data, the first 60 years data were used to calibrate the models and the remaining data were used for validation. For SVR, Gaussian radial basis function kernel was used. The first 60 years data were used for optimizing the value of SVM parameters, ε, C, and γ. A grid search was used for optimizing the values of parameters using 10-fold cross-validation. The values for the parameters have been varied in an exponentially growing sequence with C=20, 22, …, 222, γ=2−15, 2−13, …, 23 and epsilon=2−1, 20, …, 24. The values of these parameters were selected separately for all the stations. The optimized values of these three parameters for different stations are presented in Table 3.

Table 3:

Optimized Values of SVM Model Parameters for Different Stations.

StationSVR1SVR2
ε(2x)C(2x)γ(2x)ε(2x)C(2x)γ(2x)
1−12−1312−15
248−1316−15
3−110−15−14−15
4010−1304−15
502−13−14−13
6410−15−16−15
714−1116−11
846−906−15
9−12−13−12−15
10−110−1516−15
11−18−1508−15
1216−9−14−15
13−116−1504−15
1436−1516−15
1512−13−14−13
1624−13−14−15
1746−1304−15
18−16−1114−15
1902−13−14−13
2034−1334−13
21−12−1124−13

For WR, rainfall series was pre-processed using DWT for calibration and validation periods separately. In this study, Haar wavelet (db1) and five levels of decompositions were used. The WR model structure is shown in Figure 5. Using DWT, a lagged monthly rainfall time series was decomposed into subcomponents, i.e. approximation (Ai) and detail (Di) components. A new series was generated by adding the effective Di components and the fifth approximation component (A5). The effective Di were chosen based on their correlations with the original time series of rainfall. The newly generated time series was used as input to the LR model, and the original rainfall time series was used as output.

Figure 5: Wavelet Regression Model Structure.
Figure 5:

Wavelet Regression Model Structure.

4.1 Comparison of BR, SVR, and WR

Four different statistics, namely mean absolute error (MAE), root mean square error (RMSE), correlation coefficient (R), and Nash-Sutcliffe efficiency coefficient (NS) [43], were used to evaluate the performance of the three techniques. These statistics are defined as follows:

  1. The MAE between the observed and the predicted outputs can be defined as

    MAE=1ni=1n|fiyi|.

  2. The RMSE between the observed and the predicted outcomes can be defined as

    RMSE=1ni=1n(fiyi)2.

  3. R is defined as

    R=i=1n(yiy¯i)(fif¯i)[i=1n(yiy¯i)2·i=1n(fif¯i)2]12.

  4. NS is defined as

    NS=1i=1n(fiyi)2i=1n(yiy¯i)2,

where fi are the estimated values, f̅i is the average of estimated values, yi are the observed values, and y̅i is the average of observed values.

Figure 6A–D shows the comparison of the performance of six models during the validation period. As observed from the figure, the WR2 model outperformed all other models for all 21 stations. The MAE for WR2 was observed to lie between 62.89 mm at station no. 19 and 119.813 mm at station no. 11. The RMSE for WR2 was found to lie between 93.06 mm (for station no. 15) and 185.79 mm (for station no. 11), while the RMSE for SVR2 model was observed between 106.15 and 214.08 mm. The average RMSEs for BR1, SVR1, WR1, BR2, SVR2, and WR2 were 172.84, 151.24, 142.58, 173.89, 145.42, and 127.06 mm, respectively. The WR2 model had the least RMSE compared to all other models.

Figure 6: (A) MAE for Different Models at 21 Stations. (B) RMSE for Different Models at 21 Stations. (C) Correlation Coefficient (R) for Different Models at 21 Stations. (D) Nash-Sutcliffe Index (NS) for Different Models at 21 Stations.
Figure 6:

(A) MAE for Different Models at 21 Stations. (B) RMSE for Different Models at 21 Stations. (C) Correlation Coefficient (R) for Different Models at 21 Stations. (D) Nash-Sutcliffe Index (NS) for Different Models at 21 Stations.

Figure 6C shows the R for different models at all stations during the validation period. The value of R for WR2 was found to be between 0.794 at station no. 21 and 0.855 at station no. 8. The average values of R for BR1, SVR1, WR1, BR2, SVR2, and WR2 were 0.667, 0.702, 0.750, 0.643, 0.729, and 0.828, respectively. Figure 6D shows the Nash-Sutcliffe efficiency index (NSE) for different models at all stations during the validation period. The maximum and minimum values of NSE for WR2 were 0.694 at station 8 and 0.578 at station 21. The average values of NSE for BR1, SVR1, WR1, BR2, SVR2, and WR2 were 0.334, 0.491, 0.547, 0.327, 0.529, and 0.640, respectively. Considering R and NSE, it is clear that the WR2 model performed better than all other models.

Rainfall is one of the most complex and challenging hydrological processes to understand and to model because of the complexities involved with the associated atmospheric processes that generate rainfall. The rainfall time series is affected by many complex factors having different frequency components. Considering only one resolution component, the model is unable to capture the internal mechanism behind the rainfall process. Using DWT, the complex rainfall time series were decomposed into several simple time series, and the WR model was developed utilizing appropriate subcomponents. The performance of WR model was better than the other models due to the fact that it uses different resolution components for LR. The ineffective frequency components were removed from the rainfall time series, which effectively improved the performance of the LR model. WR is a simpler model compared to the non-linear SVR model; however, WR has shown better accuracy in rainfall prediction. This signifies the role of wavelet preprocessing of the rainfall time series, which de-noised the inputs and resulted in improved prediction ability of the model.

To assess the spatial variations in the performance of the forecasting models, the state was divided into four zones: lower Assam, middle Assam, upper Assam, and southern Assam (Figure 7). Table 4 presents the average RMSEs for four regions, obtained using different models. A comparison of model performances reveals that the upper Assam part has the highest accuracy (least RMSEs), whereas middle Assam has the least accuracy (highest RMSEs). The average standard deviations of monthly rainfall for upper, middle, lower, and southern Assam were 168, 254, 222, and 259 mm, respectively. The upper Assam region had the minimum average standard deviation, which could be the reason for the better performance of models in this region, as variability in station rainfalls for this region was least. Figure 7 shows the performance of the WR2 model (best model) in the form of a map made using inverse distance weighted interpolation of RMSEs at different stations. The average RMSEs for lower, middle, upper and southern Assam were 126.26, 154.33, 95.31, and 141.25 mm, respectively. Upper Assam had the lowest RMSE followed by lower, southern, and middle Assam, respectively. The WR2 model performed best for the stations in the upper Assam zone. Stations in the middle Assam had the highest RMSEs, which imply that the forecasting model performed badly at these stations. Of 21 stations, station no. 15 (Lakhimpur) had the least RMSE for the WR2 model, whereas station no. 11 (Kamrup) had the highest RMSE. It is already shown in Table 1 that the Kamrup station had the highest values of standard deviation and skewness.

Figure 7: Map Showing the Distribution of RMSEs for the WR2 Model over Assam.
Figure 7:

Map Showing the Distribution of RMSEs for the WR2 Model over Assam.

Table 4:

RMSEs Obtained for Different Parts of Assam Using Six Models.

RegionRMSE (mm)
BR1SVR1WR1BR2SVR2WR2
Upper127.42111.81106.19128.12108.7195.31
Middle206.65180.82169.10207.37176.85154.33
Lower182.18157.02147.59182.88144.55126.26
Southern192.27170.08161.48194.80163.41141.25

5 Conclusions

A comparison of three different forecasting techniques, BR, SVR, and WR, for monthly rainfall data at 21 stations in Assam, India, is presented in this paper. The GRBF kernel was used for SVR, and the SVR parameters were optimized utilizing a grid search technique. For the WR model, the rainfall time series was decomposed into subcomponents using DWT and a new time series was generated by adding effective subcomponents, which were used as input for the LR model. The monthly rainfall data of 102 years were used for the study, which were divided into 60:40 ratio for calibration and validation, respectively. Based on the MAE, RMSE, R, and NSE, WR was found to be the most efficient and BR was found the least efficient for rainfall prediction. The efficiency index for BR, SVR, and WR was found to be 32.8%, 52.9%, and 64.03%, respectively. The WR model used in this study was a linear model; however, SVR is an advanced non-linear model. WR is simpler than SVR but was able to perform better than SVR. Hence, it can be concluded that wavelet processing of the input time series resulted in better performance of the WR model. Spatial assessment of model performance reveals that model performance was found best for stations in the upper part of the state, followed by the lower, southern, and middle parts. The study can be useful for better management of water resources in the state.

Bibliography

[1] M. Behzad, K. Asghari, M. Eazi and M. Palhang, Generalization performance of support vector machines and neural networks in runoff modeling, Expert Syst. Appl.36 (2009), 7624–7629.10.1016/j.eswa.2008.09.053Suche in Google Scholar

[2] M. Bray and D. Han, Identification of support vector machines for runoff modelling, J. Hydroinform.6 (2004), 265–280.10.2166/hydro.2004.0020Suche in Google Scholar

[3] C. Chou and R. Wang, On-line estimation of unit hydrographs using the wavelet-based LMS algorithm, Hydrol. Sci. J.47 (2003), 721–738.10.1080/02626660209492976Suche in Google Scholar

[4] A. S. Cofino, R. Cano, C. Sordo and J. M. Gutierrez, Bayesian networks for probabilistic weather prediction, Proc. 15th Eur. Conf. Artif. Intell.700 (2002), 695–700.Suche in Google Scholar

[5] S. Coles and J. A. Tawn, A Bayesian analysis of extreme analysis rainfall data process, J. Roy. Stat. Soc. Ser. C45 (2011), 463–478.10.2307/2986068Suche in Google Scholar

[6] X. Dai, P. Wang and J. Chou, Multiscale characteristics of the rainy season rainfall and interdecadal decaying of summer monsoon in North China, Chin. Sci. Bull.48 (2003), 2730–2734.10.1007/BF02901765Suche in Google Scholar

[7] Y. B. Dibike, S. Velickov, D. Solomatine and M. B. Abbott, Model induction with support vector machines: introduction and applications, J. Comput. Civ. Eng.15(2001), 208–216.10.1061/(ASCE)0887-3801(2001)15:3(208)Suche in Google Scholar

[8] L. Frechette and A. Khan, Bayesian regression-based urban traffic models, Transp. Res. Rec.1644 (1998), 157–165.10.3141/1644-17Suche in Google Scholar

[9] M. K. Goyal, Monthly rainfall prediction using wavelet regression and neural network: an analysis of 1901–2002 data, Assam, India, Theor. Appl. Climatol.118 (2014), 25–34.10.1007/s00704-013-1029-3Suche in Google Scholar

[10] A. Grossmann and J. Morlet, Decomposition of Hardy functions into square integrable wavelets of constant shape, SIAM J. Math. Anal.15 (1984), 723–736.10.1137/0515056Suche in Google Scholar

[11] S. R. Gunn, Support vector machines for classification and regression, Technical Report, 1998.Suche in Google Scholar

[12] W. C. Hong and P. Ping-Feng, Potential assessment of the support vector regression technique in rainfall forecasting, Water Resour. Manag.21 (2007), 495–513.10.1007/s11269-006-9026-2Suche in Google Scholar

[13] T. Joachims, Optimizing search engines using clickthrough data, in: Proc. Eighth ACM SIGKDD Int. Conf. Knowl. Discov. Data Min. KDD’ 02, pp. 133–142, 2002.10.1145/775047.775067Suche in Google Scholar

[14] K. Kim, Financial time series forecasting using support vector machines, Neurocomputing55 (2003), 307–319.10.1016/S0925-2312(03)00372-2Suche in Google Scholar

[15] S. U. Kim and K. S. Lee, Regional low flow frequency analysis using Bayesian regression and prediction at ungauged catchment in Korea, KSCE J. Civ. Eng.14 (2009), 87–98.10.1007/978-3-540-89465-0_57Suche in Google Scholar

[16] O. Kisi, Wavelet regression model as an alternative to neural networks for monthly streamflow forecasting, Hydrol. Process.23 (2009), 3583–3597.10.1002/hyp.7461Suche in Google Scholar

[17] O. Kisi, Wavelet regression model as an alternative to neural networks for river stage forecasting, Water Resour. Manag.25 (2011), 579–600.10.1007/s11269-010-9715-8Suche in Google Scholar

[18] D. Labat, R. Ababou and A. Mangin, Rainfall-runoff relations for karstic springs, part II: Continuous wavelet and discrete orthogonal multiresolution analyses, J. Hydrol.238 (2000), 149–178.10.1016/S0022-1694(00)00322-XSuche in Google Scholar

[19] S. Lee, S. Cho and P. M. Wong, Rainfall prediction using artificial neural networks, J. Geogr. Inf. Decis. Anal.2 (1998), 233–242.Suche in Google Scholar

[20] J. Y. Lin, C. T. Cheng and K. W. Chau, Using support vector machines for long-term discharge prediction, Hydrol. Sci. J.51 (2006), 599–612.10.1623/hysj.51.4.599Suche in Google Scholar

[21] S. G. Mallat, A theory for multiresolution signal decomposition: the wavelet representation, IEEE Trans. Pattern Anal. Mach. Intell.11 (1989), 674–693.10.1109/34.192463Suche in Google Scholar

[22] H. D. Navone and H. A. Ceccatto, Predicting Indian monsoon rainfall: a neural network approach, Clim. Dyn.10 (1994), 305–312.10.1007/BF00228029Suche in Google Scholar

[23] V. B. Nikam and B. B. Meshram, Modeling Rainfall Prediction Using Data Mining Method: A Bayesian Approach, pp. 132–136, IEEE, 2013.10.1109/CIMSim.2013.29Suche in Google Scholar

[24] E. Osuna and R. Freund, An improved training algorithm for support vector machines, in: Neural Networks Signal Process VII – Proc. IEEE, pp. 276–285, 1997.Suche in Google Scholar

[25] S. Raghavendra and P. C. Deka, Support vector machine applications in the field of hydrology: a review, Appl. Soft Comput. J.19 (2014), 372–386.10.1016/j.asoc.2014.02.002Suche in Google Scholar

[26] O. A. Rosso, A. Figliola, S. Blanco and P. M. Jacovkis, Signal separation with almost periodic components: a wavelets based method, Rev. Mex. Fis.50 (2004), 179–186.Suche in Google Scholar

[27] N. Sethi and K. Garg, Exploiting data mining technique for rainfall prediction, Int. J. Comput. Sci. Inf. Technol.5 (2014), 3982–3984.Suche in Google Scholar

[28] A. Sharma and M. K. Goyal, Bayesian network model for monthly rainfall forecast, in: 2015 IEEE International Conference on Research in Computational Intelligence and Communication Networks (ICRCICN), Kolkata, India, pp. 241–246, IEEE, 2015.10.1109/ICRCICN.2015.7434243Suche in Google Scholar

[29] N. Sharma, M. Zakaullah, H. Tiwari and D. Kumar, Runoff and sediment yield modeling using ANN and support vector machines: a case study from Nepal watershed, Model Earth Syst. Environ.1 (2015), 23.10.1007/s40808-015-0027-0Suche in Google Scholar

[30] C. Sivapragasam, S. Liong and M. Pasha, Rainfall and runoff forecasting with SSA-SVM approach, J. Hydroinform.3 (2001), 141–152.10.2166/hydro.2001.0014Suche in Google Scholar

[31] L. Smith, D. Turcotte and B. Isacks, Stream flow characterization and feature detection using a discrete wavelet transform, Hydrol. Process.12 (1998), 233–249.10.1002/(SICI)1099-1085(199802)12:2<233::AID-HYP573>3.0.CO;2-3Suche in Google Scholar

[32] A. M. Ticlavilca and M. McKee, Multivariate Bayesian regression approach to forecast releases from a system of multiple reservoirs, Water Resour. Manag.25 (2011), 523–543.10.1007/s11269-010-9712-ySuche in Google Scholar

[33] V. N. Vapnik, The nature of statistical learning theory, Springer, Berlin, 1995.10.1007/978-1-4757-2440-0Suche in Google Scholar

[34] V. N. Vapnik, An overview of statistical learning theory, IEEE Trans. Neural Netw.10 (1999), 988–999.10.1109/72.788640Suche in Google Scholar

[35] V. Vapnik, S. E. Golowich and A. Smola, Support vector method for function approximation, regression estimation, and signal processing, in: Advances in Neural Information Processing Systems, vol. 9, San Mateo, CA, 1996.Suche in Google Scholar

[36] W. Wang and J. Ding, Wavelet network model and its application to the prediction of hydrology, Nat. Sci. Sleep, 1 (2003), 67–71.Suche in Google Scholar

[37] C. C. Wei, Wavelet support vector machines for forecasting precipitation in tropical cyclones: comparisons with GSVM, regression, and MM5, Weather Forecast27 (2012), 438–450.10.1175/WAF-D-11-00004.1Suche in Google Scholar

[38] J. P. Liu, M. Q. Chang, X. Y. Ma, Groundwater quality assessment based on support vector machine. In: HAIHE River Basin Research and Planning Approach-Proceedings of 2009 International Symposium of HAIHE Basin Integrated Water and Environment Management, Beijing, China, 2009, pp. 173–178.Suche in Google Scholar

[39] Z. Zakaria and A. Shabri, Streamflow forecasting at ungaged sites using support vector machines, Appl. Math. Sci.6 (2012), 3003–3014.Suche in Google Scholar

[40] H. C. Zhou, Y. Peng and G. H. Liang, The research of monthly discharge predictor-corrector model based on wavelet decomposition, Water Resour. Manag.22 (2008), 217–227.10.1007/s11269-006-9152-xSuche in Google Scholar

[41] A. Mahanta, D. Jhingran, C. K. Das and A. Partie. Assam Human Development Report 2003, 2003, Available at http://hdr.undp.org/en/reports/nationalreports/asiathepacific/india/name,3268,en.html. Accessed 20 July, 2013.Suche in Google Scholar

[42] G. Urcid and G. X. Ritter, Advances in knowledge-based and intelligent information and engineering systems, in: Advances in Knowledge-Based and Intelligent Information and Engineering Systems, M. Graña, C. Toro, J. Posada, R. J. Howlett and L. C. Jain, eds., pp. 2140–2149, IOS Press, 2012.Suche in Google Scholar

[43] J. E. Nash and J. V. Sutcliffe, River flow forecasting through conceptual models, part I – a discussion of principles, J. Hydrol.10(1970), 282–290.10.1016/0022-1694(70)90255-6Suche in Google Scholar

Received: 2016-4-13
Published Online: 2016-9-17
Published in Print: 2017-9-26

©2017 Walter de Gruyter GmbH, Berlin/Boston

This article is distributed under the terms of the Creative Commons Attribution Non-Commercial License, which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Heruntergeladen am 25.9.2025 von https://www.degruyterbrill.com/document/doi/10.1515/jisys-2016-0065/html
Button zum nach oben scrollen