1. INTRODUCTION
One of the most prominent issues in discussing the economy of a region is the issue of economic growth. Economic growth is the development of activities in the economy that cause goods and services produced in society to increase causing the welfare of society increases as well (Imansyah et al., 2017;Putratama et al., 2012). Economic growth is considered important in the context of the economy because it can be a measurement of economic achievement in the area. GRDP is an indicator that can be used to measure the economic growth of an area (Mbarek and Feki, 2013;Meltem et al., 2012). A yeartoyear value of GRDP can give an idea of the achievement of regional economic performance in those years. Therefore, future GRDP values also need to be foreseen to know how regional economic developments are going forward (Saman, 2011;Vrbka, 2016;Singh et al., 2018, Jalali and Jaafar, 2019).
The data of GRDP value movement follows a time series based on the order of data points within a certain time interval. Time series analysis can be applied to predict the probability of future circumstances. One time series model can be used in forecasting is ARIMA. The principle of the ARIMA model is to predict the value of observed data in the future based on a series of observations made in the past (Ili et al., 2016;Suwardo et al., 2010;Kodekova et al., 2018). In addition to the ARIMA model, Artificial Neural Networks (ANN) is also one of the methods that can use in forecasting. ANN is an artificial representation of the human brain that continuously learns to solve a problem. One of the learning methods in ANN is Backpropagation. In the learning process, ANN forms a model based on the pattern of training results of some data. LevenbergMarquardt is one of the training algorithms that can accelerate learning from ANN (Claveria and Torra, 2013;Haider and Hanif, 2009;Jaramillo, 2018, Asad et al., 2018).
Researches on forecasting using time series and ANN analysis have done (Ramli et al., 2018). Forecast the inflation in Ghana with the ANN method, using the data of monthly inflation from January 1991 to December 2010. The objective is to forecast and project inflation from January 2011 to December 2011. The results of ANN also compared with the time series model such as Autoregression AR (12) and Vector Autoregression VAR (14), which uses the same set of data variables. The basis of comparison is an outofsample estimated error (RMSFE). The results show that RMSFE of ANN is lower than that of the time series model. That is, with these prediction comparison predictions based on the ANN model are more accurate.
Sukono et al. (2016), researched the modelling and forecasting of inflation data in Indonesia using time series models and feedforward neural networks. The purpose of the research is to develop ANN modelling in statistical structure and apply it to inflation data in Indonesia. The study also aims to optimise the development of the FFNN procedure model and its application. The FFNN model and time series model are evaluated based on Root Mean Square Error (RMSE), Mean Absolute Deviation (MAD) and MAPE. The best model criteria are those that have a minimum of RMSE, MAD, and MAPE. The result of comparative analysis between FFNN model and time series model shows that the FFNN model is better than the time series model.
Chaudhuri and Ghosh (2016) were researching the application of ANN model and time series in the forecasting exchange rate of a currency. To forecast exchange rates, used two different classes of approaches, which based on ANN model and time series model. Multilayer Feed Forward Neural Network (MLFFNN) and Nonlinear Autoregressive model with Exogen Neural Network Input (NARX) is an approach used as an ANN model. Generalised Autoregressive Conditional Heteroscedastic (GARCH) and Exponential Generalized Autoregressive Conditional Heteroskedastic (EGARCH) used as time series models. The results show that although two different approaches are efficient in forecasting exchange rates, MLFNN and NARX are the most efficient.
Based on these, this paper focuses on research on how to predict GRDP Bandung Regency Indonesia, using Artificial Neural Network Backpropagation Levenberg Marquardt method. The goal is to form a model forecast of GRDP Bandung Regency Indonesia, using the ARIMA model, and the Artificial Neural Network Backpropagation LevenbergMarquardt method, as well as to compare which of the approaches is better.
2. MATERIALS AND METHODS
This section discusses the object of research and the stages of research data analysis which includes: bootstrap resampling, ARIMA models, and models of Artificial Neural Networks (ANN) Backpropagation Levenberg Marquardt method.
2.1 Bootstrap Resampling
According to Timmerman et al. (2009), in doing research, sometimes the sample data obtained is much less than the population. These causes the result of sample data processing to be less representative in giving a conclusion. The bootstrap method was introduced by Efron in 1993, with the aim of obtaining the best estimates based on minimal data with the help of computers. Bootstrap is a method of resampling the sample data with the condition of return on the data that can represent the actual population data. Bootstrap resampling can be done with the help of software R. However, raw data resampling using R software cannot be used directly in this research. The data used in this research is new data made based on the average value and standard deviation of the overall resampling result data.
2.2 ARIMA Models
There are two important representations in describing a time series data pattern, i.e. AR and Moving Average (MA) models. However, a new problem arises when one of the models used contains too many parameters (order is too large). The number of parameters that are too much can reduce the accuracy of the model in forecasting. To overcome these, there can be a combination of both AR and MA models into the ARMA model, which is a stationary time series model (Sukono et al., 2017). When p declares the order of the AR model, and q declares the order of the MA model, then the ARMA (p, q) model is expressed as follows:
with X_{t} random variable GRDP at time $t;{\phi}_{1},\mathrm{...},{\phi}_{p}$ parameters of the AR model; ${\theta}_{1},\hspace{0.17em}\mathrm{...},\text{}\hspace{0.17em}{\theta}_{q}$ parameters of the MA model, and a_{t} residual value at time t.
The ARIMA model is a nonstationary time series model denoted by ARIMA (p, d, q), where p declares the order of the AR model, q declares the order of the MA model, and d states the differencing order (Cadenas et al., 2016). The ARIMA model stated as follows:
where in AR operation ${\phi}_{p}(B)=(1{\phi}_{1}B{\phi}_{2}B\mathrm{...}{\phi}_{p}{B}^{p})$ and MA operation ${\theta}_{q}(B)=(1{\theta}_{1}B{\theta}_{2}B\mathrm{...}{\theta}_{q}{B}^{q})$, with X_{t} random variable GRDP at the time t; B backshift operator; ${(1B)}^{d}{X}_{t}$ time series is stationary on differencing to d; and a_{t} residual value at the time t.
The steps of time series modelling are as follows: (i) Identification of the model, by determining the order value of p and q by autocorrelation function (ACF) and partial autocorrelation function (PACF) plot. (ii) Parameter Estimation by least squares method or maximum likelihood. (iii) Diagnostic test by white noise test and noncorrelation serially against residual a_{t}. (iv) Forecasting, by deciding if the model is suitable so that it used for recursive prediction.
2.3 Artificial Neural Network
ANN is an artificial representation of the human brain that is continuously learning to solve a problem (Nakamura, 2005;Rădulescu and Banica, 2014). ANN mimics the workings of human biological neural networks in processing information. The elements involved in this process are interconnected, and they operate in parallel. Similar to human neural networks, ANN is also composed of a set of neurons. In the process of sending information between neurons, these neurons connected by one or more networks that have weight values. The weight value obtained through the learning process, where one form of learning method ANN is Backpropagation. The simple form of neurons interpreted in Figure 1.
In ANN, neurons are located and scattered in layers called neuron layers. There are three layers of the ANN constituent neurons: (i) Input Layer; the input neurons are in this layer. The input neuron receives input from outside. The input is a description of a problem. (ii) Hidden Layers; Hidden neurons are in this layer. The output of the hidden layer cannot observed directly. (iii) Output Layer; The output neurons are in this layer. The output of this layer is the output of artificial neural networks in the form of a solution to a problem.
2.3.1 Activation Function
The activation function is a function used in the processing of input and bias values to produce the output value of neurons (Mohammed et al., 2015;Ozbek et al., 2011). The following are two commonly used activation functions (equation 3). These Linear Functions; the linear function has an output value equal to its input value. The linear function formulated as follows:
Binary sigmoid function; Binary sigmoid function has value in the range (0, 1). Therefore, this function used very commonly in ANN which usually produces output at intervals between 0 and 1. The binary sigmoid function formulated:
This function has a derivative:
2.3.2 Backpropagation Model
Backpropagation is one of the learning methods of artificial neural networks. Backpropagation trains the network to strike a balance between the network's ability to recognise patterns during training with the network's ability to respond appropriately to the patterns used during the training (Naik and Pathan, 2012;Qi and Zhang, 2008). The activation function used in backpropagation is a binary sigmoid function. Backpropagation training divided into three phases: Phase I, Propagation Forward (feedforward); in this phase, the input signal x_{i} is propagated (counted forward) by using the activation function to generate network output y_{k}. These output results will be comparing with the expected target. The difference between the two is called an error. If the error is smaller than the specified tolerance, then iteration is stopped. However, if the opposite happens, then the weights on each line must be changed to minimise errors. Phase II, Backpropagation; in this phase, the error occurring is propagated backwards from the line directly related to the units in the output layer. Distribution error from output to the previous layer made by using delta function and Phase III, Change of Weight; in this phase, the weights of all layers are modified simultaneously. This change based on the delta neuron in its upper layer. The three phases of this training are done repeatedly until the conditions of termination of the iteration fulfilled.
In the Backpropagation training phase, a condition is needed to stop the training process and as a benchmark for network accuracy in recognising the given pattern. One of the error calculations that used as a benchmark in the termination of training and the measurement of accuracy is the Mean Squared Error (MSE). MSE is the mean error of the square of the difference between network output and target output. The purpose of this training process is to get as minimal error value as possible by replacing the value of weights connected to all neurons on the network. The MSE formulated as follows:
with k: input to k; t_{k}: target output signal to k; and y_{k}: output signal for network prediction to k.
2.3.3 LevenbergMarquardt Method
The LevenbergMarquardt algorithm is one type of ANN Backpropagation training algorithm. LevenbergMarquardt is one solution to improve the performance of Backpropagation because the algorithm is designed to minimize the error sum of squares. The algorithm of Backpropagation LevenbergMarquardt is described as follows:

Step 1: Initialize weights and biases with small random values and maximum epoch;

Step 2: Determining the required parameters, among others: (i) Initialization epoch = 0, (ii) LevenbergMarquardt Parameters μ whose value must be greater than zero; and (iii) Factor parameters Beta β which is used as a parameter multiplied or divided by the LevenbergMarquardt parameter;

Step 3: For each pair of training data, if epoch <epoch maximum or MSE> target error, do the following steps.
Phase I: Propagation Forward (Feedforward)

Step 4: Each input unit x_{i}, i = 1, 2, ..., n receives input signal xi and passes them to each unit in the hidden layer;

Step 5: Each hidden unit z_{i}, j = 1, 2, ..., p sums the weighted input signal v_{ij} and bias vo_{j}, by formula:
Use the activation function to calculate the output signal:
Next, the signal z_{j} will be sent to all units on the layer above it (output layer).

Step 6: Each unit of output y_{k}, $k=1,\hspace{0.17em}2,\hspace{0.17em}\mathrm{...},\hspace{0.17em}m$ sums the weighted input signal w_{jk} and bias ${w}_{{0}_{k}}$, with the formula
Use the activation function to calculate the output signal:
Next, the signal y_{k} will be sending to all units on the layer above it.
Phase II: Reverse Propagation (Backpropagation)

Step 7: Calculating error and MSE; the formula for calculating error is:
where k = output to k, so that MSE can be calculated using the formula:
The formula to calculate the total error: $e={({e}_{1}\text{}{e}_{2}\text{}\mathrm{...}\text{}{e}_{n})}^{T}$

Step 8: Calculates an error on each unit of the output layer y_{k}, $k=1,2,\hspace{0.17em}\mathrm{...},\hspace{0.17em}m$ with the formula:

Step 9: Calculates errors on each hidden layer unit by ${z}_{j},\hspace{0.17em}j=1,2,\hspace{0.17em}\mathrm{...},p$ summing its input delta:
So the error information:
$${\delta}_{{1}_{j}}=\delta \_i{n}_{j}\times f\text{'}(z\_i{n}_{j}\times [{z}_{j}(1{z}_{j})],$$(17) 
Step 10: Fforming the matrix J
Phase III: Change Weight

Step 11: Calculate new weights
with:

Step 12: Counting MSE
If $MS{E}_{new}\le MS{E}_{old}$ then,

Step 13: Training stops when epoch = epoch is up.
2.3.4 Normalisation and Denormalization of Data
Normalisation was done with the aim to achieve a certain range using the input data and targets during the training phase and ANN testing. Normalisation minmax scales a given value into a new value between 0 and 1, where the scaling based on the maximum and minimum value of the data set. The formula in normalising the data is as follows:
with X_{n} normal data values; X actual data values; X_{min} the minimum value of all actual data; and X_{max} the maximum value of the actual data. Meanwhile, denormalisation of data is the process of returning data to the initial form before the normalisation. The following is the formula of denormalisation:
with X_{n} normal data values; X actual data values; X_{min} the minimum value of all actual data; and X_{min} the maximum value of the actual data.
2.3.5 Evaluate Error in Forecasting
The difference between the forecast result and the actual data can be expressing in the Mean Absolute Percentage Error (MAPE), which formulated as follows:
with n the amount of data; x_{t} actual data on time t; and ${\widehat{x}}_{t}$ output of forecast results at a time t. If the MAPE value is less than 10%, then the forecasting result is said to be very good. While if the value of MAPE is between 1020%, then the forecasting results are said to be good.
3. RESULT AND ANALYSIS
This section discusses the result and analysis which includes: GRDP data of Bandung Regency Indonesia, data analysis using time series of ARIMA model, data analysis using ANN Backpropagation LevenbergMarquardt method, and comparative analysis.
3.1 GRDP Data of Bandung Regency, Indonesia
GRDP data used in this study obtained from the official site of the Central Bureau of Statistics Bandung Regency Indonesia. The data taken is the GRDP data at constant prices based on the field of business from 2010 to 2016. The graph of the data series GRDP Bandung Regency Indonesia can be seen in Figure 2.
Based on the plot in Figure 2, it can be seen that Bandung Regency GRDP has increased from year to year in 20102016. These means that the economy of Bandung Regency also progressed from year to year in 20102016. However, the available GRDP data is still too small to be processed because time series analysis requires much data in order to obtain a predictive model. Therefore, the data needs to be resampled using bootstrap. The data resampled 400 times, and the result of plot resampling of Bandung Regency GRDP data as shown in Figure 2.
3.2 Data Analysis Using the ARIMA Time Series Model
Before entering the analysis phase, 400 GRDP data of Bandung Regency resampling converted to GRDP data from September 1983 to December 2016.
3.2.1 Stationary Test
The stationary test needs to be done, to know stationary GRDP data. The stationary test has done with Augmented DickeyFuller (ADF) test, with the help of Eviews 9.0 software. Hypothesis used is H_{0}: data is not stationary, and H_{1}: data is stationary. From the test results, a statistical value obtained is t_{statistic} with a GRDP data of t_{statistic} = 20.22669, while the critical value of statistics at a significance level of 5% is t_{critical} = 2.868565. Therefore, t_{statistic} > t_{critical} is applied, so the hypothesis of H_{0} the ADF test is rejected at 5% significance level, meaning that with 95% confidence the GRDP data has been stationary.
3.2.2 Model Identification
Identification of time series models can be made with Autocorrelation Function (ACF) and Partial Autocorrelation Function (PACF) plots, using Eviews 9.0 software. At level data, ACF and PACF plots do not indicate the order of a model, in the absence of a cutoff lag. Therefore, to overcome this problem, GRDP data is done in a different degree of one. Thus, the time series used is a model of ARIMA (p, 1, q). Furthermore, the plots of ACF and PACF of GRDP data of different degree of one can already indicate the order of a time series model. The ACF plot, cuts off significantly in lag 1, while the PACF plot decreases exponentially from lag 1. So there is an indication to model the GRDP data which can be done using ARIMA (0, 1, 1) model. However, time series modelling is also tried using ARIMA (1, 1, 1) and ARIMA (1, 1, 0) models.
3.2.3 Selection of the Best Model
In the previous model identification stage, it has said that threetime series models, i.e. ARIMA (0, 1, 1) models, ARIMA (1, 1, 1), and ARIMA (1, 1, 0) are attempted to considered. Choosing the best model can be done by using the Akaike Information Criterion (AIC), where the smallest AIC value shows the best model. Based on GRDP data processing with Eviews 9.0 software, obtained AIC value from each model is as follows: ARIMA(0, 1, 1) model with value of AIC = 32.59592; ARIMA (1, 1, 1) with value of AIC = 32.60034; and ARIMA(1, 1, 0) with value of AIC = 33.02344. The smallest AIC value is 32.59592 generated by the ARIMA (0, 1, 1) model, which makes the best model.
3.2.4 Parameter Estimation
Parameter estimation performed on ARIMA (0, 1, 1) tentative model by using Eviews 9.0 software. Estimation result of ARIMA (0, 1, 1) tentative parameter model first stage was obtained by constant estimator c of 1015,685, and parameter estimator for the coefficient of a_{t1} is 0.992332. However, the estimated value of the constant c of 1015,685 was not significant at a significance level of 5%. It is, therefore, necessary to estimate the second stage parameter, by not including the constants parameter c. Parameter estimators obtained the second stage estimation results for coefficients a_{t1} of 0.991942. Referring to equation (1), estimator model ARIMA (0, 1, 1) for Bandung Regency GRDP is
3.2.5 Parameter Significance Test (Test statistics t)
Against the model estimator (27) it is necessary to test the significance of the parameters, using statistical tests t. In the significance test at equation (27) only the parameter estimator for the coefficients was performed a_{t1}. The test hypothesis used is H_{0}: θ_{1} = 0. Means parameter coefficient of a_{t1} had no significant effect on X_{t}, and H_{t}: θ_{1} ≠ 0, whereas means parameter coefficient a_{t1} had a significant effect on X_{t}. The results of tests conducted with the help of Eviews 9.0 software obtained statistical parameters for coefficients a_{t1} amount t_{statistic} = 267.9172. Meanwhile, at 5% significance level, the statistical critical value of t_{(0.025; 4002)} = 1.96, so t_{statistic} > t_{(0.025; 4002)} was applied. Hence, hypothesis H_{0} was rejected at a 5% significance level, meaning that with the confidence of 95%, parameter estimator for coefficients a_{t1} has a significant effect on X_{t}. Figure 3
3.2.6 Test Validity (Test statistics F)
The timing model estimate (27) also needs to be validated by statistical test F, with the aim of measuring the effect of parameters collectively in the model. The hypothesis used is H_{0}: ϕ_{1} = θ_{1} = 0, meaning that all parameters in the model have significant influence together with X_{t}, and ${H}_{1}:\exists {\phi}_{1}\ne {\theta}_{1}\ne 0$ meaning that there is at least one parameter that has a significant effect on X_{t}. Validation testing is done with the help of Eviews 9.0 software, with a generated value of F_{statistic} = 405.1586. While at a significance level of 5%, the critical value of statistic F_{(0.05; 2; 40021)} = 3, so F_{statistic} > F_{(0.05; 2; 40021)} was applied. Hence, the hypothesis H_{0} is rejected at a 5% significance level. These means that with a confidence of 95% there is at least one parameter that has a significant effect on X_{t}.
3.2.7 Diagnostic Test
Once the model parameter estimator has been influenced significantly, further diagnostic testing of the residual model is necessary, in order to ensure that the model estimator is feasible for use in forecasting. Diagnostic testing is performed with two treatments, i.e. assumption of residual normality, and white noise test. Residual Normality Test, conducted with the aim to determine the residual on the model of normal distribution. The hypothesis used is H_{0}: a_{t} normally distributed, and H_{t}: at not normally distributed. The residual normality test on the model was performed using the JarqueBera (JB) statistical test, with the help of Eviews 9.0 software. Descriptive statistics of Bandung Regency GRDP, include 400 resampling data, with skewness of 0.202645 and kurtosis of 3.411359, based on JarqueBera formula obtained value of JB = 5.55787. JarqueBera statistics are asymptotic to the ChiSquare distribution. Using the 5% significance level and the degrees of freedom 2, statistically critical values obtained were ${\chi}_{(0.05;2)}^{2}=5.99$, so $JB<{\chi}_{(0.05;2)}^{2}$, was applied. Thus hypothesis H_{0} was received at a 5% significance level, meaning that with the confidence of 95% residual a_{t} in ARIMA (0, 1, 1) the model estimator is a normal distribution.
White Noise test is conducted to determine the serial correlation between lag on residual. The hypothesis used is ${H}_{0}:{\rho}_{1}=\mathrm{...}={\rho}_{m}=0$, meaning there is no residual correlation between lag, and ${H}_{1}:\exists {\rho}_{k}\ne 0$ for a $k\in \left\{1,\mathrm{...},m\right\}$, meaning that there is a residual correlation between lag. White noise testing was performed using the LjungBox Q (m) test statistic where m is the degrees of freedom, through the residual correlogram plot of ARIMA (0, 1, 1) model, with the help of Eviews 9.0 software. Based on the results of the correlogram plot, it appears that all lag does not pass through the Bartlett line, and the value Pr [Q (m)] for every lag m with m = 1, 2,.., 20, has a greater value than the level of significance $\alpha =0.05$. Therefore, it concluded that with 95% confidence, no residual correlation occurs between lag and residual in the model, which means that the ARIMA (0, 1, 1) model follows the white noise process.
3.2.8 Forecasting
After passing many tests of parameter significance, validation test, and diagnostic test, the best estimator of the ARIMA (0, 1, 1) model was obtained in the equation as follows:
This model can used for forecasting Bandung Regency GRDP. The comparison between forecasting results data and actual data of Bandung Regency GRDP, during September 1983  December 2016 can be seen in the graph in Figure 4
Figure 4 above shows that the data forecast of Bandung Regency GRDP approached the actual data. The degree of difference or error occurring between these two data can be calculated using the Mean Absolute Percentage Error (MAPE) with equation (26), and a MAPE value of 3.90% obtained.
3.2.9 Data Normalization
Before performing data analysis stage using ANN Backpropagation LevenbergMarquardt method, the data first of all needs to be normalised. These is because the activation function used is a binary sigmoid function, where it has a range value between 0 and 1. Normalisation of data done by applying equation (24). After data normalisation done, it was then used for the design of network structure, as follows.
3.3 Data Analysis Using ANN Backpropagation LevenbergMarquardt Method
This section discusses data analysis using ANN Backpropagation LevenbergMarquardt method, which includes: data normalisation, network structure design, training, and network testing, optimal network selection, data forecasting, and denomination.
3.3.1 Data Normalization
Before performing data analysis stage using ANN Backpropagation LevenbergMarquardt method, the data first of all needs to be normalised. These is because the activation function used is a binary sigmoid function, where it has a range value between 0 and 1. Normalisation of data done by applying equation (24). After data normalisation done, it was then used for the design of network structure, as follows.
3.3.2 Designing Network Structure
After the data was normalised, the next step was to initialise the parameters used in ANN. These parameters can be shown in Table 1.
Based on the parameter data in Table 1, there are five different combinations of inputs used in the Backpropagation process:

Input on lag1 (X_{t}_{1})

Input on lag1 (X_{t}_{1}) and lag 2 (X_{t}_{2})

Input on lag1 (X_{t}_{1}), lag 2 (X_{t}_{2}), and lag 3 (X_{t}_{3})

Input on lag1 (X_{t}_{1}), lag 2 (X_{t}_{1}), lag 3 (X_{t}_{3}), and lag 4 (X_{t}_{4})

Input on lag1 (X_{t}_{1}), lag 2 (X_{t}_{2}), lag 3 (X_{t}_{3}), lag 4 (X_{t}_{4}), and lag 5 (X_{t}_{5})
A trial and error method conducted for each input, with some neurons between 1 to 5 to obtain optimal network in producing 1 GRDP data output. Networks were trained using LevenbergMarquardt (trainlm) by applying the binary sigmoid activation function.
3.3.3 Training and Testing Network
Backpropagation attempts to strike a balance between the recognition of proper training patterns, and a good response to other similar patterns. Therefore, the data are divided into two parts, the data pattern used as training, and the data used as the test. In this research, the training phase is done by using 280 GRDP data of Bandung Regency, whereas, in the testing phase, the data used were as many as 80 GRDP data. The remaining data used in the validation process. Afterwards, the data went through the training process, and network testing was done using the help toolbox in the MATLAB software. The results of the training process and backpropagation testing on the GRDP data using the LevenbergMarquardt training method with one lag1 input (X_{t}_{1}), in generating one output with the number of neurons in the hidden layer between 1 and 5, is shown in Table 2.
Table 2 shows the best selected FFNN (FFNN) Backpropagation or Feed Forward Neural Network (FFNN) model is the least likely to is rated MSE testing compared to other FFNN models. This FFNN model consists of 1 input, five neurons, and one output. The value of MSE FFNN (1, 5, 1) for the training stage is 0.0165924, and testing stages 0.0117982. Then, the results of the training process and Backpropagation testing on GRDP data, using the LevenbergMarquardt training method with two inputs on lag1(X_{t}_{1}) and lag 2 (X_{t}_{2}) in producing one output with the number of neurons in the hidden layer between 1 and 5, is shown in Table 3.
Table 3 depicts the best selected FFNN model is FFNN (2, 4, 1) because it has the smallest MSE testing value compared to other FFNN models. This FFNN model consists of 2 inputs, four neurons, and one output. The MSE value in the FFNN (2, 4, 1) model for the training stage is 0.0165924, whereas testing stage is 0.0157151. Furthermore, the results of the training process and Backpropagation testing on the PDRB data using the LevenbergMarquardt training method with three inputs on lag 1(X_{t}_{1}), lag 2 (X_{t}_{2}), and lag 3 (X_{t}_{3}) in producing 1 output with the number of neurons in the hidden layer between 1 and 5, is shown in Table 4.
Table 4 demonstrates that the best selected FFNN model is FFNN (3, 4, 1) because it has the smallest MSE testing value compared to other FFNN models. This FFNN model consists of 3 inputs, four neurons, and one output. The MSE value in the FFNN (3, 4, 1) model for the training stage is 0.0158606, whereas testing stage is 0.0135612. The results of the training process and Backpropagation testing on GRDP data using the LevenbergMarquardt training method with four inputs on lag1 (X_{t}_{1}), lag 2 (X_{t}_{2}), lag 3 (X_{t}_{3}), and lag 4 (X_{t}_{4}) in generating 1 output with the number of neurons in the hidden layer between 1 and 5, is shown in Table 5.
Table 5 indicates that the best selected FFNN model is FFNN (4.5.1) because it has the smallest MSE testing value compared to other FFNN models. This FFNN model consists of 4 inputs, five neurons, and one output. The MSE value in the FFNN (4.5.1) model for the training stage is 0.0165779, whereas testing stage is 0.0175038. Finally, the results of the training process and Backpropagation testing on GRDP data using the LevenbergMarquardt training method with five inputs on lag 1 (X_{t}_{1}), lag 2 (X_{t}_{2}) , lag 3 (X_{t}_{3}), lag 4 (X_{t}_{4}), and lag 5 (X_{t}_{5}) in producing 1 output with the number of neurons in the hidden layer between 1 and 5, is shown in Table 6.
Table 6 shows that the best selected FFNN model is FFNN (5, 2, 1) because it has the smallest MSE testing value compared to other FFNN models. This FFNN model consists of 5 inputs, two neurons, and one output. The MSE value in the FFNN (5, 2, 1) model for the training stage is 0.0166623, whereas testing stage is 0.015937.
3.3.4 Selection of Optimal Network
After training and testing the network using Backpropagation LevenbergMarquardt is done, the best FFNN model from each combination of input and hidden neurons are obtained. The comparison of MSE values from these best FFNN models can be shown in Table 7.
Table 7 shows the FFNN (1,5,1) model is the FFNN model which has the smallest MSE testing value compared to other FFNN model that is 0.0117982. Therefore, it can be concluding that FFNN (1.5.1) is the model with the most optimal network in predicting the GRDP data of Bandung Regency.
3.3.5 Forecasting and Denormalization of Data
The data of network output or forecast result using FFNN (1, 5, 1), then needed to go through denormalisation using equation (25). The objective is to return it to its original form before normalisation, in order to make a real difference to the actual data of GRDP. After the data is normalised, the comparison graph between the network output data and the actual GRDP data is formed as shown in Figure 5.
Figure 5 displays the difference in network output data, and actual data of GRDP value is not much different. Data output network is the data whose value is relatively constant. The degree of difference or error between these two data can be calculated using MAPE with equation (26). Based on the calculation a MAPE value of 3.88% was obtained.
3.4 Comparative Analysis of ARIMA and ANN Backpropagation Model LevenbergMarquardt Method
To find out which model is better to be used in predicting the GRDP of Bandung Regency, both the ARIMA model and ANN Backpropagation LevenbergMarquardt methods need to go through comparative analysis. Based on model estimation analysis that has been done, obtained MAPE for model ARIMA (0, 1, 1) is equal to 3.90%, while MAPE for FFNN (1, 5, 1) is equal to 3.88%. Because the MAPE value of FFNN (1, 5, 1) model is smaller, compared to MAPE ARIMA (0, 1, 1) model, it can be concluded that FFNN (1, 5, 1) model is a better model used in forecasting the GRDP of Bandung Regency. However, the MAPE values obtained from the ARIMA and FFNN models in this study are limited to the scope of 400 GRDP datasets of Bandung Regency resampling results using bootstrap. Therefore, it is possible to obtain different results if the ARIMA and FFNN models applied to other data.
4 CONCLUSIONS
In this paper, we have analysed the model of GRDP forecasting using ANN Backpropagation LevenbergMarquardt method. Based on the results of the analysis, it can be concluding as follows: the best ARIMA time series model in forecasting of Bandung Regency GRDP is ARIMA (0, 1, 1). With MAPE error rate in of 3.90%, and the best model of ANN Backpropagation LevenbergMarquardt method for forecasting of Bandung Regency GRDP is FFNN (1, 5, 1), having one input on lag1, five hidden neurons, and one output, with MAPE error rate of 3.88%. Based on the MAPE error value of both models, a better model used in predicting of Bandung Regency GRDP is ANN Backpropagation LevenbergMarquardt method, with the FFNN (1, 5, 1) model.