Neural Network Intelligent Learning Algorithms for ...

18 downloads 17873 Views 731KB Size Report
energy product price prediction, including other domains, whereas comparing the learning. 77 algorithms are limited despite its significance in turning the NNs ...
CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 1

Neural Network Intelligent Learning Algorithms for Inter-Related Energy

2

Products Applications

3

Haruna Chiroma1, Sameem Abdulkareem1, Sanah Abdullahi Muaz2, Abdullah Khan3, Eka Novita Sari4, and Tutut Herawan4

4 5 6 7

1

Department of Artificial Intelligence Department of Software Engineering 4 Department of Information systems University of Malaya 50603 Pantai Valley, Kuala Lumpur, Malaysia 2 Department of Information Systems, International Islamic University Malaysia, Kuala Lumpur, Malaysia 3 Software and Multimedia Centre Universiti Tun Hussein Onn Malaysia 86400 Parit Raja, Batu Pahat, Johor Darul Takzim, Malaysia 2

8 9 10 11 12 13 14 15 16 17 18 19 20

[email protected], [email protected], {sameem,tutut}@um.edu.my, [email protected],

21 *Corresponding author, e-mail: Haruna Chiroma, University of Malaya, [email protected], [email protected]

22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37

ABSTRACT: Accurate prediction of energy products future price is required for effective reduction of future price uncertainty as well as risk management. Neural Networks (NNs) are an alternative to statistical and mathematical methods of predicting energy product prices. The daily prices of Propane (PPN), Kerosene Type Jet fuel (KTJF), Heating oil (HTO), New York Gasoline (NYGSL), and US Coast Gasoline (USCGSL) interrelated energy products are predicted. The energy products prices are found to be significantly correlated at 0.01 level (2tailed). In this paper, NNs learning algorithms are used to build a model for the accurate prediction of the five (5) energy products prices. The aptitudes of the five (5) NNs learning algorithms in the prediction of PPN, KTJF, HTO, NYGSL, and USCGSL are examined and their performances are compared. The five (5) NNs learning algorithms are Gradient Decent with Adaptive learning rate backpropagation (GDANN), Bayesian Regularization (BRNN), Scale Conjugate Gradient backpropagation (SCGNN), Batch training with weight and bias learning rules (BNN), and Levenberg-Marquardt (LMNN). Simulated comparative results suggest that LMNN and BRNN can be viewed as the best NNs learning algorithms in terms of R 2 and MSE 1

Corresponding author: Haruna Chiroma, Department of Artificial Intelligence, University of Malaya. 50603 Pantai Valley, Kuala Lumpur, Malaysia Email: [email protected], Tel: +60143873685

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 38 39 40 41 42

whereas GDANN was found to be the fastest. This study has provided an alternative approach to the prediction of energy products prices, which can reduce the high level of uncertainty about energy products prices. Therefore, provide a platform for developmental planning that can result in the improvement of economic standard.

43

KEYWORDS: US Coast Gasoline, Heating oil, Propane, Bayesian Regularization, Levenberg-

44

Marquardt

45 46

INTRODUCTION

47 48

The future prices of energy products such as Propane (PPN), Kerosene Type Jet fuel (KTJF),

49

Heating oil (HTO), New York Gasoline (NYGSL), and US Coast Gasoline (USCGSL) are

50

highly uncertain. The uncertainty trailing these energy products prices has succeeded in

51

attracting both domestic and foreign political attention, and this facilitated market ranking [1].

52

Accurate forecasting of future prices of energy product can effectively be used for risk

53

management as argued by [2]. Malliaris and G. Malliaris [2] forecast one month ahead spot

54

prices of crude oil, heating oil, gasoline, natural gas and propane since their spot prices in market

55

are interrelated. Spot price data of crude oil, heating oil, gasoline, natural gas, and propane

56

collected from Barchart (www.barchart.com) for a period starting from January 3,1994 to

57

December 31, 2002. The data that cover December 1997 – November 2002 were used as

58

experimental sample data for building the forecasting models. Multi linear regression, Neural

59

Networks (NNs) model, and simple model were applied in each of the energy market to forecast

60

one month future prices of the energy product. Results show the NNs perform better than the

61

statistical models in all markets except for propane market. Wang and Yang [3] examined the

62

probability of predicting crude oil, heating oil, gasoline, and natural gas futures markets within a

63

day by conducting experiments with different models: NN, semi parametric function coefficient,

64

nonparametric kernel regression, and generalized autoregressive conditional heteroskedasticity

65

with data collected for 30 minutes Intraday prices and returns of the four energy future contract

66

source from NYMEX. For each of the four individual future contracts, 15 – 20 years prices of

67

future contracts were analyzed, a period where the price was low and in steady decline (bear

68

market), another period where the price is high and on a steady increase (bull market) were

69

identified. After performing experiments with the collected data and employed models, results

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 70

indicated only heating oil and natural markets possessed the possibility of being predicted within

71

a day especially under bull market condition. The NNs was found to outperform the statistical

72

models. However, these statistical methods assume normal distribution for input data [4] which

73

makes the statistical methods unsuitable for energy products price prediction because of the non-

74

linear, complex and volatile nature of the energy products, experimental evidence can be found

75

in [5-6]. Therefore, the comparison of NNs and statistical methods might not provide a fair

76

platform. Most literature mainly focuses on comparing the architecture of NNs in the domain of

77

energy product price prediction, including other domains, whereas comparing the learning

78

algorithms are limited despite its significance in turning the NNs weights and bias. In this study

79

we have chosen a multilayer NN learning algorithms because recurrent NN structure becomes

80

more complex, thus, further complicates the chosen of the best NN parameters; the computation

81

of the error gradient in a recurrent NN architecture also turn out to be complicated due to

82

presents of more attractors in the state space of a recurrent NN [7].

83

In this paper, we propose to evaluate and compare the validity of fast NNs learning algorithms as

84

a useful technique for the prediction of energy products price. The NNs learning algorithms are

85

used to build a model for the prediction of PPN, KTJF, HTO, NYGSL, and USCGSL prices.

86

Subsequently, compare the performances of the learning algorithms in each of the market.

87 88

BACKGROUND

89 90

In the literature, several studies were conducted on the comparison of prediction performances of

91

different NNs architecture. Gencay and Liu [8] Compare the performances of feed forward

92

neural network (FFNN) and recurrent neural network (RNN). Support vector machine (SVM)

93

and back propagation neural network (BPNN) were contrasted in [9 – 10]. The RNN and FFNN

94

were compared in a study conducted by [11]. Time delay (TDNN), RNN and probabilistic neural

95

networks (PNN) were compared by [12]. The study of performance comparison between BPNN

96

and SVM is reported in [13]. FFNN, RNN and Elman recurrent network (ERN) were compared

97

in [14]. Also, performances of SVM and BPNN were compared in [15]. ANFIS, FFNN and

98

radial basis function networks (RBFN) compared in [16]. ERN, FFNN and ANFIS comparative

99

studies are presented in [17]. Conventional support vector machine (CSVM) and improved

100

cluster support vector machine (ICSVM) contrasted in [18]. Comparative studies among BPNN,

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 101

SVM regression (SVMR) and RBFN were performed by [19]. The FFNN and SVM

102

performances were contrasted in [20]. Finally, a study in [21] compared SVMR and RBFN. The

103

summary of the results obtained from the comparative studies are reported in Table 1.

104

Table 1 Comparing performance accuracy of NNs architectures

105 106 Reference [8]

[9-10]

[11]

Result

Domain

RNN outperforms FFNN

Signal processing

SVM outperform PBNN

Finance

FFNN perform better than RNN

Crude oil price

RNN perform better than TDNN, and PNN

Stock trading

[12] First case (BPNN perform better than SVM), [13]

Second case (SVM outperform BPNN)

Crude oil price and Natural gas

RNN perform better than FFNN, and ERN

Option trading and hedging

SVM outperform BPNN

Crude oil price

ANFIS perform better than FFNN and RBFN

Natural Gas

ANFIS perform better than ENN, and FFNN

Crude oil

CSVM perform better than the CSVM

Crude oil

[14]

[15]

[16]

[17]

[18]

First case (RBFN perform better than SVM and BPNN), second case (SVM outperform RBFN, and

Crude oil

BPNN) [19]

[20]

SVM outperform BPNN

Drugs

SVMR perform better than RBFNN

Rainfall

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. [21]

107 108

Table 1 reported established results in the literature showing different NNs architectures which

109

are used to build a model and prediction results generated by the models were compared to

110

assess performance accuracy. Table 1 clearly showed no specific NNs architecture is suitable

111

across all problem domains.

112 113 114 115 116

MATERIALS AND METHODS

117 118

Neural networks learning algorithms

119

The weights and bias of NNs are iteratively modified during NNs training to minimize error

120

function such as Eqn.(1):

121

MSE 

122

2 1 N   x ( j )  y ( j )  N j 1

(1)

123 124 125

Where N, x (j), and y (j) are the total number of predictions made by the model, original

126

observation in the dataset, and the value predicted by the model, respectively. The closer the

127

value of MSE to zero (0), the better is the prediction accuracy but do not mean the zero (0) MSE

128

that typically occure due to overfitting). Zero (0) indicates a perfect prediction, which rarely

129

occurs in practice. The most widely use a NN learning algorithm is the BP algorithm which is a

130

gradient-descent technique of minimizing an error function. The synaptic weight (W) in a BP

131

learning algorithm can be updated using Eqn. (2):

Wk 1  Wk  Wk

132

(2)

133

Here, k is the iteration in a discrete time and the current weight adaptation is represented by Wk

134

expressed as:

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing.

Wk  

135

ek Wk

(3)

ek are learning rate (typically ranges from 0 to 1) and gradient of the error Wk

136

Where  and

137

function to be minimized, respectively. The main drawbacks of the gradient descent BP includes:

138

slow convergence speed and possibility of being trapped in local minima as a result of its

139

iterative nature of solving problem till the error function reaches its minimal level. Appropriate

140

specification of learning rate and momentum determine the success of BP in a large scale

141

problem. Gradient-decent BP is still being applied in many NNs programs. Though, the BP is no

142

more considered as the optimal and efficient learning algorithm. Thus, powerful learning

143

algorithms that are fast in convergence are developed based on heuristic method from the

144

standard steepest descent algorithm referred to as the first category of the fastest learning

145

algorithms. The second category of the fastest learning algorithms was developed based on

146

standard numerical optimization methods such as the Levenberg-Marquardt (LM). Typically,

147

conjugate gradient algorithms converge faster than the variable learning rate BP algorithm, but

148

such results are limited to an application domain, implying that the results can differ from one

149

problem domain to another different domain. The conjugate gradient algorithms require line

150

search for each iteration, which makes the conjugate gradient to be computationally expensive.

151

The scaled conjugate gradient backpropagation (SCGNN) algorithm was developed in response

152

to the computationally expensive nature of the conjugate gradient so as to speed up convergence.

153

Other alternative learning algorithms includes Gradient Decent with Adaptive learning rate

154

backpropagation (GDANN), Bayesian Regularization (BRNN), Batch training with weight and

155

bias learning rules (BNN), and Levenberg-Marquardt (LMNN). However, LMNN is viewed as

156

the most effective learning algorithm for training a medium sized NNs. Gradient descent is used

157

by the LM to improve on its starting guess for tuning the LMNN parameters [22-23].

158

Energy product dataset and descriptive statistics

159

The daily spot prices of HTO, PPN, KTJF, USCGSL, and NYGSL were collected from 9 July,

160

1992 to 16 October, 2012 source from the Energy Information Administration of the US

161

Department of Energy. The data were freely available, published by the Energy Information

162

Administration of the US Department of Energy. The data were collected on daily basis since

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 163

enough data are required for building a robust NNs model. The data comprised of five thousand

164

and ninety (5090) rows and four (4) columns. The data were not normalized to prevent the

165

destruction of the original pattern in the historical data [24]. The descriptive statistics of the data

166

are computed and the results are reported in Table 2. The standard (Std.) Deviation displayed in

167

the last column of Table 2 indicated uniform dispersion among the energy products prices.

168 169

Table 2 Descriptive Statistics of energy products, datasets

170 Product

N

Min

Max

Mean

Std. D.

PPN

5090

0.2

1.98

0.6969

0.40872

KTJF

5090

0.28

4.81

1.037

0.73589

HTO

5090

0.28

4.08

1.2622

0.88739

NYGSL

5090

0.29

3.67

1.2472

0.82894

USCGSL

5090

0.27

4.87

1.2298

0.82616

171 172

Table 3 is a correlation among the energy product prices. The correlation is significant among

173

the HTO, PPN, KTJF, USCGSL, and NYGSL as clearly showed in Table 3. Correlated variables

174

imply that influence on a variable can affect the other variable positively as points out in Table 3.

175

Hair et al. [25] argued that for better prediction, variables in the research data have to be

176

significantly correlated. Therefore, HTO, PPN, KTJF, USCGSL can be independent variables

177

whereas NYGSL dependent variable. Also, PPN, KTJF, USCGSL, and NYGSL can be used as

178

independent variables whereas HTO dependent variable. This can also be applied to PPN, KTJF,

179

and USCGSL. Therefore, we compared the NNs learning algorithms in five different energy

180

markets based on the datasets.

181 182

Table 3 Correlation matrix of the energy products datasets

183 PPN

184

KTJF

HTO

NYGSL

0.984**

0.997**

**

KTJF

0.702

HTO

0.949**

0.745**

NYGSL

0.937

**

0.737**

USCGSL

0.940**

0.727**

** Correlation is significant at the 0.01 level (2-tailed).

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 185 186

Neural network modeling

187

After several trials, our data were partition into training, validation, and test with 3562, 764, and

188

764 samples, respectively. To avoid over-fitting the training data, random sampling was used to

189

partition the dataset. Updating of NNs weights and bias as well as computation of the gradient is

190

performed with the training dataset. To explore the best combination of the activation functions

191

(ACFs), several ACFs are considered: log-sigmoid, linear, soft max, hyperbolic tangent sigmoid,

192

triangular basis, inverse and hard-limit. The hidden layer was tried with each of the ACFs while

193

linear ACFs were constantly maintained in the input layer. In the output layer linear is used to

194

avoid limiting the values in a particular range. Therefore, both input and output layers used linear

195

ACFs throughout the training period. Momentum and learning rate were varied between zero (0)

196

to one (1). The single has hidden layer is used since [26]: Theorem For every continuous non

197

constant function  every r and every pr,bility measure µ on R r ,  r ,

198

Mr, where  is a probability meAirure taken convenience to describe the relative frequency

199

oRRoccurrence of inputs  ,  r is the bored field of Rr and Mr is the set of all bored measurable

200

functions from Rr to R. Different experimental trials were performed to find the appropriate NNs

201

model with the best MSE, R2, and convergence speed. The training terminates after six (6)

202

iterations without performance improvement to avoid over-fitting the network. The network

203

architecture with the minimum MSE, highest R2, and low convergence speed are saved as

204

optimal NNs topology. The predictive capabilities of the NNs learning algorithms are evaluated

205

on test dataset.



  ( x) is  dense in r

206 207

RESULTS AND DISCUSSION

208 209

The proposed algorithms were implemented in MATLAB 2013a Neural Network ToolBox on a

210

computer system (HP L1750 model, 4 GB RAM, 232.4 GB HDD, 32-bit OS, Intel (R) Core

211

(TM) 2 Duo CPU @ 3.00 GHz). The number of hidden neurons should not be twice of the

212

independent variables as argued in [27]. Thus, we consider between four (4) to ten (10) ranges of

213

the number of neurons and used to verify for NNs optimal architecture for every learning

214

algorithms. Different ACFs were experimentally tried with corresponding number of hidden

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 215

neurons. The models with the best results are reported in Tables 4 to 8 and these with poor

216

results are discarded. The best ACFs found for the prediction of LMNN is log-sigmoid, for BNN

217

is a hyperbolic tangent sigmoid, for GDANN is log-sigmoid, for SCGNN is triangular basis and

218

for BRNN is log-sigmoid. The probable reason for having different ACF for the separate

219

architecture can be attributed to the inconsistent behavior of the NNs architecture. Tables 4 to 8

220

shows performance (Mean Square Error (MSE)) (Regression (R2)) and convergence speed

221

(Iterations (I) (Time (T) in seconds (Sec.)) for each of the NNs learning algorithms. The

222

momentum and learning rate found to be optimal were 0.3 and 0.6, respectively. The minimum

223

MSE, highest R2, optimum combinations of I and T are in bold throughout the Tables.

224 225

Table 4 Performance of the prediction of HTO price with different NNs learning algorithms

226 Performance Learning Method

LMNN

SCGNN

GDANN

BNN

BRNN

Convergence speed

Number of hidden neurons

Number of hidden neurons

4-8-1

4-7-1

4-6-1

4-8-1

4-7-1

4-6-1

0.000178

0.000115

0.00082

65

(0.9959)

(0.9938)

(0.9955)

(5)

71(5)

72(4)

0.00208

0.00504

0.00363

1000

1000

1000

(0.8790)

(0.9808)

(0.9246)

(150)

(117)

(100)

2.86

3.93

0.793

(-0.8471)

(-0.9142)

(0.7940)

1(0)

1(0)

1(0)

8.89

5.57

8.63

1000

1000

1000

(-0.7898)

(-0.6456)

(0.9410)

(27)

(27)

(26)

0.00573

0.00635

0.00669

217

230

293

(0.9963)

(0.9961)

(0.9958)

(25)

(22)

(22)

227 228

From Table 4 it can be deduced that the LMNN algorithm has the lowest MSE and is the fastest

229

in converging to the optimal MSE whereas BRNN achieved better R2 (see Fig.1a ) than the other

230

learning algorithms. These results indicated that the performance of the algorithms in predicting

231

HTO is not consistent because it depends on the performance metrics being considered as the

232

criteria for measuring performance. Though, in this case LMNN can be chosen despite not

233

having the highest R2 due to its ability to achieve the lowest MSE in the shortest possible time.

234

Seven (7) hidden neurons produce the best MSE result, whereas six (6) hidden neurons is the

235

fastest architecture.

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 236

Table 5 Comparison of KTJF predicted by NNs learning algorithms models

237 238

Learning Method

LMNN

Performance

Convergence speed

Number of hidden neurons

Number of hidden neurons

4-8-1

4-7-1

4-6-1

4-8-1

4-7-1

4-6-1

0.102(0.8771)

0.0873(0.9245)

0.1(0.88)

47(5)

110(9)

54(2)

1000

1000

1000

3.2(-0.67)

(183)

(168)

(142)

0.29(0.72)

12(2)

16(2)

31(10)

1000

1000

1000

(28)

(27)

(28)

322

531

443

(40)

(49)

(33)

SCGNN

1.3000(-0.4111)

34.3000(-0.7662)

GDANN

0.62(0.7482)

0.744(-0.7108)

BNN

BRNN

15.7(-0.5972)

29(0.7396)

0.0896(0.9108)

13.6(0.68)

0.0896(0.91468)

0.098(0.9)

239 240

The results of the prediction of KTJF price are reported in Table 5, the LMNN has the minimum

241

MSE and the highest R2 (see Fig.1b) among the comparison algorithms. The other algorithms

242

such as BRNN, GDANN (4-8-1, 4-6-1) also have competitive values of MSE and R2 compared

243

to the optimal values. The fastest algorithm is GDANN having the minimum iterations and time

244

of convergence. The performance criteria’s indicated that the LMNN is the best in terms of MSE

245

and R2 whereas convergence speed criteria’s shows GDANN outperforms other algorithms.

246

Seven (7) hidden neurons yield the best MSE and R2 but, the fastest architecture is having eight

247

(8) hidden neurons. This is surprising as less complex structure is expected to be the fastest.

248 249 250 251

Table 6 The comparison of the NNs learning algorithms in the prediction of NYGSL

252 Performance

Convergence speed

Number of hidden neurons

Number of hidden neurons

Learning Method 4-8-1

4-7-1

4-6-1

4-8-1

4-7-1

4-6-1

LMNN

0.0044(0.9983)

0.00158(0.9988)

0.00195(0.9987)

39(3)

60(3)

35(1)

SCGNN

1(-0.1631)

18.6(0.4774)

13.5(0.9531)

1000(183)

1000(153)

1000(141)

GDANN

3.22(0.8825)

2.12(0.8311)

0.490(0.9471)

1(0)

1(0)

1(0)

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. BNN

13(-0.5668)

5.22(0.9338)

0.864(0.8711)

BRNN

0.000025(0.9989)

0.0000279(0.9988)

1000(27)

1000(25)

1000(24)

0.000512(0.9988) 433(48) Training: R=0.91249

954(87)

604(44)

Validation: R=0. 4

Data Fit Y=T

Output ~= 0.83*Target + 0.17

4

Output ~= 0.84*Target + 0.16

253 254

3.5 R=0.99637 price are reported Test: R=0.99568GDANN as the The results of the predictionTraining: of NYGSL in Table 6, showing 4 3

255

Fit 3.5 Fitlearning algorithm with2.5 fastest algorithm to converge to its optimal solution. The BRNN 3.5 2.5

256

2 different architecture is3 the best predictor with the lowest MSE and the highest R2 (see Fig.1c) 2

257

among the comparison 2algorithms.

2.5

1.5 1

1.5

0.5 1

1

3

3

Training: R=0.99904 Data Fit Y=T

20.5 1

2

1.5

3

4

Target

b

1

4.5

Data Fit Training: Y=T

4 3.5 1.8 3 1.6 2.5 1.4 2 1.2 1.5

1

2

Target 1

1.5

2

3

3.5

0.8 0.5 1

2

4.5

R=0.9846

Data Fit Y=T

3 2.5

3

4

0.2 0.5

Data Fit Y=T

4 3.5 3 2.5

Validation Data Fit Y=T

1.8 1.6 1.4

2

1.2

1.5

1

1 0.5

2

Target 1.5

0.4

4

All: R=0.9144

Test: R=0.99845

1 1

0.6

2.5

Target

0.8 0.6 1

2

3

Target

0.4 0.2

1

1

1.5

0.5

Target

Ta

0.5 1

1.5

2

2.5

3

3.5

0.5

Target

Predicted PPN by LMNN

All: R=0.99895 3.5

Data Fit Y=T

3 2.5 2 1.5 1 0.5

Data Fit Y=T

1.8 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2

1

1.5

2

2.5

3

3.5

1

Target

Target

d

1.5

2

2.5

3

Target

All: R Data Fit Y=T

1.8 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2

0.5 0.5

1

Test: R=0.98502

Output ~= 0.97*Target + 0.021

0.5

Predicted NYGSL by BRNN

1

4

Data Fit Y=T

0.5

c

3

0.5

LMNN

Data Fit Y=T

2.5 1

a

1.5

0.5 1

Test: R=0.92453

+ 0.021 ~= 0.97*Target OutputPredicted by KTJF

+ 0.0024 Output ~= 1*Target HTO predicted

by BRNN

4 3.5

31.5

1.5

4

Output ~= 1*Target + 0.002

2

All: R=0.99627

2

2

2

Target

2.5

2.5

0.5

1

3.5

Y=T

3

Target

0.5

260

Data

Output ~= 0.83*Target + 0.17

259

Y=T

3

Output ~= 0.96*Target + 0.026

258

Data

Output ~= 0.99*Target + 0.0099

Output ~= 0.99*Target + 0.0092

4

Data Fit Y=T

3.5

1.5

0.5

1

Ta

Output ~=

Output ~=

2 1.5

2 1.5

Predicted USCGSL by BRNN

1 CITATION: Chiroma, H., Abdul-Kareem,1 S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. 0.5 0.5 (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products 1 2 3 4 1 2 3 Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Target Target Publishing.

All: R=0.99825 Data Fit Y=T

4.5 4 3.5 3 2.5 2 1.5 1 0.5 1

2

3

4

Target

e 261

Fig. 1 Regression plots

262 263 264 265 266

SCGNN and GDANN are having the poorest values of MSE despite GDANN has a competitive

267

R2 compared to the promising R2 value of BRNN, LMNN, BNN (4-7-1, 4-6-1), and SCGNN (4-

268

6-1). In the prediction of NYGSL, the performance exhibited a similar phenomenon to the

269

prediction of HTO and KTJF as consistency is not maintained. In the prediction of the NYGSL

270

we cannot conclude on the best algorithms because the performance exhibited by the algorithms

271

is highly random unlike the case in the prediction of HTO. The BRNN converged to the MSE

272

and R2 very slow compared to the LMNN, and GDANN speed. The optimal algorithm in this

273

situation depends on the criteria chosen as the priority in selecting the best predictor. If accuracy

274

is the priority, then BRNN can be the best candidate, whereas speed place LMNN above BRNN.

275

Seven (7) hidden neurons have the best MSE value, whereas the architectures with six (6), seven

276

(7), and eight (8) hidden neurons are the fastest. This could probably be caused by memorizing

277

the training data by the algorithms.

278 279

Table 7 Results ontain with NNs learning algorithms in the prediction of PPN

280 Performance Learning Method

LMNN

Convergence speed

Number of hidden neurons

Number of hidden neurons

4-8-1

4-7-1

4-6-1

4-8-1

4-7-1

4-6-1

0.00514(0.9850)

0.00694(0.9801)

0.00623(0.9820)

74(8)

32(2)

60(4)

4

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 1000

1000

1000

SCGNN

0.266(-0.4121)

0.646(-0.8601)

0.686(-0.8592)

(179)

(149)

(136)

GDANN

1.78 (-0.8164)

2.45 (-0.2453)

3.31 (0.7886)

1(0)

1(0)

1(0)

1000

1000

1000

BNN

0.276(-0.28011)

0.211(0.3802)

1.62(-0.4091)

(26)

(25)

(25)

BRNN

0.000512(0.984)

0.00557(0.98231)

0.686(-0.8592)

204(19)

226(21)

385(28)

281 282 283

Table 7 indicated that GDANN is the fastest to predict PPN price, whereas the MSE of BRNN is the

284

best. The R2 (see Fig.1d) value of LMNN is the highest compared to the SCGNN, GDANN, BNN, and

285

BRNN R2 values. The performance exhibited by the algorithms in the prediction of PPN price is not

286

different from that of NYGSL, KTJF, and HTO because consistent performance is not realized. The best

287

algorithm for the prediction of PPN price depends on the performance metrics considered as priority for

288

selecting the optimal algorithm as earlier explained. The algorithms with negative values of R2 reported

289

in Tables 4 to 7 suggested that the observed price and the predicted once are in opposite directions.

290

Signifying that upward movement of predicted price can influence the observed price to move

291

downward and vice-versa. This is not true considering the promising results obtained by other

292

algorithms that show positive R2 values.

293

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing.

Table 8 Comparison of USGSL predicted by NNs learning algorithms models

294 295

Performance

Convergence speed

Number of hidden neurons

Number of hidden neurons

Learning Method 4-8-1

4-7-1

4-6-1

4-8-1

4-7-1

4-6-1

LMNN

0.00264(0.99821)

0.00330(0.99817)

0.00301(0.99624)

103(7)

21(1)

63(3)

SCGNN

0.00598(0.99686)

0.00403(0.99621)

0.00445(0.99818)

37(1)

146(4)

131(1)

GDANN

0.0531(0.95758)

0.0284(0.97383)

0.0484(0.96635)

55(0)

68(0)

94(1)

BNN

0.0246(0.98191)

0.712(0.69757)

0.913(0.96988)

1000(23)

10(0)

12(0)

BRNN

0.00250(0.99825)

0.00273(0.99785)

0.00235(0.94193)

652(67)

653(57)

563(41)

296 297

In the prediction of USCGSL price as indicated in Table 8, BRNN have the minimum value of

298

MSE and the highest R2 (see Fig.1e), though with a different hidden neurons. The fastest

299

algorithm is the GDANN with seven (7) hidden neurons. It seems hidden layer neurons do not

300

always affect the convergence speed of the NNs algorithms based on experimental evidence

301

from Tables 4 to 8. The results do not deviate from similar behavior shown by the prediction of

302

HTO, PPN, NYGSL, and KTJN prices.

303

Small number of iterations do not necessarily imply lower computational time based on evidence

304

from the simulation results. For example, in Table 8, a GDANN converge to a solution in sixty

305

eight (68) iterations, 0 Sec. Whereas Table 5 indicated that convergence occurs in thirty five (35)

306

iterations, one (1) Sec. with the LMNN which is considered as the most efficient NN learning

307

algorithm in the literature. The poor performance exhibited by some algorithms can be attributed

308

to the possibility that the algorithms could have been trapped in local minima. The complexity of

309

an NNs affects convergence speed as reported in Tables 4 to 8. The fastest architectures have six

310

(6) hidden neurons with exception in the prediction of KTJN. This is a multi tasking experiments

311

perform on the related energy products. We have found from the series of the experiments

312

conducted that the LMNN and BRNN constitute an alternative approaches for predition in the oil

313

market especially when accuracy is the subject of concern. Objective of the research have being

314

achieved since the idle NN learning algorithms were identified for future prediction of the energy

315

products. Therefore, uncertainty related to the oil market can be reduce to the tolerance level

316

which in turn might stabilized the energy product market. The results obtained do not agree with

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 317

the results reported by [3]. This could probably be attributed to the fair comparison of our study

318

unlike the study by [3] that compared NN and statiscal methods. The results of this study cannot

319

be generalized to other multi-task problems because the performance of NNs learning algorithm

320

depends on the application domain since the NNs performance differ from domain to domain as

321

argued by [28]. However, the methodolopgy can be modified to applied on similar datasets or

322

problem.

323

CONCLUSIONS

324

In this paper, the performance of NNs learning algorithms in energy products price prediction

325

were studied and their performances in terms of MSE, R2, and convergence speed were

326

compared. BRNN was found to have the best result in the prediction of an HTO price in terms of

327

R2 whereas LMNN achieved the minimum MSE and converges faster than the SCGNN,

328

GDANN, BRNN, and BNN in predicting HTO price. In the prediction of KTJF price, LMNN

329

performs better than the SCGNN, GDANN, BRNN, and BNN considers MSE and R 2 as

330

performance criteria’s. In contrast, GDANN is the algorithm to converge faster than the other

331

NNs learning algorithms. On the other hand, prediction of NYGSL is more effective with BRNN

332

in terms of MSE and R2, but GDANN is the fastest. BRNN have the minimum MSE whereas

333

LMNN achieved the maximum R2 in the prediction of PPN price. The fastest among the learning

334

algorithms in the prediction of PPN price is GDANN despite having the poorest MSE values.

335

BRNN performs better than the SCGNN, GDANN, LMNN, and BNN in the prediction of an

336

USCGSL price in terms of MSE and R2. GDANN recorded the best convergence speed

337

compared to SCGNN, BRNN, LMNN, and BNN.

338

The NNs learning algorithms use for the prediction of energy products prices is not meant to

339

replace the financial experts in the energy sector. Perhaps, is to facilitate accurate decision to be

340

taken by decision makers in order to reach better resolutions that could yield profits for the

341

organization. Investors in the energy sector could rely on our study to suggest future prices of the

342

energy products. This can reduce the high level of uncertainty about energy products prices,

343

thereby provide a platform for developmental planning that can result in the improvement of

344

economic standard.

345

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 346 347

REFERENCES

348 349 350 351 352 353 354 355 356

1. US Department of Energy (2004) Annual Energy Outlook 2004 with Projections to 2025. http://www.eia.doe.gov/oiaf/ archive/aeo04/index.html, Acessed 2014. 2. Malliaris ME, Malliaris SG (2008) Forecasting energy product prices. Eur J Financ 14(6) ,453 – 468. 3. Su F,Wu W (2000) Design and testing of a genetic algorithm neural network in the assessment of gait patterns. Med Eng Phys 22, 67–74. 4. Wang T, Yang J (2010) Nonlinearity and intraday efficiency test on energy future markets. Energ Econ 32, 496 – 503.

357

5. Chen X, Qu Y (2011) A prediction method of crude oil output based on artificial neural

358

networks. In: Proceeding of IEEE International Conference on Computation and Information

359

sciences, Chengdu, China, pp. 702 – 704.

360 361 362 363 364 365 366 367 368 369 370 371 372

6. Shambora WE, Rossiter R (2007) Are there exploitable inefficiencies in the futures market for oil?. Energ Econ 29,18 – 27. 7. Blanco A, Delgado M, Pegalajar MC (2001) A real-coded genetic algorithm for training recurrent neural networks. Neural Networks 14, 93 – 105. 8. Gencay R, Liu T (1997) Nonlinear modeling and prediction with feedforward and recurrent networks. Physica D.108, 119 – 134. 9. Cao L, Tay FEH (2001) Financial forecasting using support vector machines. Neural Comput 10, 184 – 192. 10. Tay FEH, Cao L (2001) Application of support vector machines in financial time series forecasting. Omega 29, 309 – 317. 11. Kulkarni S, Haidar I (2009) Forecasting model for crude oil price using artificial

neural

networks and commodity future prices. Int J Comp Sci Inf Secur 2(1), 81 – 88. 12. Saad WE, Prokhorov VDC, Wunsch CD (1998) Comparative Study of Stock Trend

373

Prediction Using Time Delay, Recurrent and Probabilistic Neural

374

Neural Networ 9(6), 1456 – 1470.

Networks.

IEEE

T

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 375

13. Fernandez V (2006) Forecasting crude oil and natural gas spot prices by classification

376

methods. IDEAS web

377

http://www.webmanager.cl/prontus_cea/cea_2006/site/asocfile/ASOCFILE12006112810582

378

0.pdf. (2006) , Access 4 January, 2014.

379 380 381 382 383 384 385 386

14. Quek C, Pasquier M, Kumar N (2008) A novel recurrent neural network-based prediction system for option trading and hedging. Appl Intell 29, 138 – 151. 15. Xie W, Yu L, Xu S, Wang S (2006) A new method for crude oil price forecasting

based

on support vector machines. Lect Notes Comput Sci 3994, 441–451. 16. Kaynar O, Yilmaz I, Demirkoparan F (2011) Forecasting of natural gas consumption with neural networks and neuro fuzzy system. Energy Educ Sci Tech 26(2), 221 – 238. 17. Zimberg B (2008) Crude oil price forecasting with ANFIS. University of Vaasa [Online]: web.http://lipas.uwasa.fi/~phelo/ICIL2008TelAviv/24.pdf. Accessed 3 July 2013.

387

18. Qi Y, Zhang W (2009) The improved SVM for forecasting the fluctuation of international

388

crude oil prices. In: Proceedings of IEEE International Conference on Electronic Commerce

389

and Business Intelligence,pp. 269 – 271.

390

19. Yu L, Wang S, Wen B (2008) An AI-agent trapezoidal fuzzy ensemble forecasting model for

391

crude oil prediction. In: proceeding of the IEEE 3rd International Conference on Innovative

392

Computing Information and Control, Dalian, China, doi:10.1109/ICICIC, 124.

393

20. Byvatov E, Fechner U, Sadowski J, Schneider G (2003) Comparison of support vector

394

machine and artificial neural networks for drug/Nondrug classification. J Chem Inf Comput

395

Sci 43, 1882-1889.

396 397 398 399

21. Chen S, Yu P, Liu B (2011). Comparison of neural network architectures and inputs for radar rainfall adjustment for typhoon events. J Hydrol 405, 150–160. 22. Zweiri YH, Whidborne JF, Sceviratne LD (2002) A three-term backpropagation algorithm. Neurocomputing 50, 305–318.

400

23. Haykin S (1999) Neural networks. New Jersey: Prentice Hall, 2nd edn.

401

24. Jammazi R, Aloui C (2012) Crude oil forecasting: experimental evidence from wavelet

402 403 404

decomposition and neural network modeling. Energ Econ 34, 828 – 841. 25. Hair FJ, Black WC, Babin JB, Anderson RE (2010) Multivariate data analysis. New Jersey: Pearson Prentice Hall.

CITATION: Chiroma, H., Abdul-Kareem, S., Muaz, S. A., Khan, A., Sari, E. N., & Herawan, T. (2014). Neural Network Intelligent Learning Algorithm for Inter-related Energy Products Applications. In Advances in Swarm Intelligence (pp. 284-293). Springer International Publishing. 405 406

26. Hornik K (1991) Approximation capabilities of multilayer feedforward networks. Neural Netw 4, 251—257.

407

27. Berry MJA, Linoff G (1997) Data mining techniques. New York: Willey.

408

28. Azar AT (2013) Fast neural network learning algorithms for medical applications. Neural

409 410

Comp Appl 23(3-4), 1019-1034.