Explosion prediction of oil gas using SVM and Logistic Regression

0 downloads 0 Views 361KB Size Report
Nov 8, 2012 - In the accidents of dangerous chemicals, the oil gas explosion plays an important role. The essential task of the explosion prevention is to ...
Explosion prediction of oil gas using SVM and Logistic Regression Xiaofei Wanga , Mingming Zhanga,b , Liyong Shena,∗, Suixiang Gaoa

arXiv:1211.1526v2 [cs.CE] 8 Nov 2012

a

School of Mathematical Sciences, University of CAS, Beijing, China b School of Engineering, University of CAS, Beijing, 100049, China

Abstract The prevention of dangerous chemical accidents is a primary problem of industrial manufacturing. In the accidents of dangerous chemicals, the oil gas explosion plays an important role. The essential task of the explosion prevention is to estimate the better explosion limit of a given oil gas. In this paper, Support Vector Machines (SVM) and Logistic Regression (LR) are used to predict the explosion of oil gas. LR can get the explicit probability formula of explosion, and the explosive range of the concentrations of oil gas according to the concentration of oxygen. Meanwhile, SVM gives higher accuracy of prediction. Furthermore, considering the practical requirements, the effects of penalty parameter on the distribution of two types of errors are discussed. Keywords: Explosion prediction, oil gas, SVM, Logistic Regression, penalty parameter

1. Introduction In recent years, there were frequent occurrences of dangerous chemical accidents in industrial processes. They bring huge damages to people’s life and property. How to predict the occurrence of dangerous chemical accidents becomes current hot topic [1, 2]. In 2006, James I. Chang and Cheng-Chung Lin [3] summarized all factory accidents in the 40 years before 2006, and classified them according to the causes of accidents. Among them, about 74% accidents occurred at oil refining plants, oil port and oil tank. The number of accidents increase year by year. Since explosions of oil tanks may cause dominoes effects which will strengthen the damages of accidents [4], it is necessary to predict and prevent the explosion of oil tank effectively. The occurrences of fire and explosion accidents (include explosions of oil tank) need three conditions, that’s, supply of oxygen, existence of combustibles and sources of kindling. If one element is controlled, combustion is impossible. According to the principles above, people naturally used to control the sources of kindling to prevent occurrences of fire/explosion accidents in the past. But since 1969, explosions of three large oil tanks happened in succession. By deep investigation and study, it comes to the conclusion that all above ∗

Corresponding author Email address: [email protected] (Liyong Shen)

Preprint submitted to Elsevier

November 9, 2012

explosions were caused by static electricity in washing cabin [5]. From then on, people get to know a new kind of kindling-static electricity, which can hardly be removed or controlled. Then people turned to study to prevent explosion by controlling another element- oxygen. As is well known, the explosion occurs only when the concentration ratio of hydrocarbon and oxygen reaches to some certain range. Recently, there have been a lot of methods to obtain the range of explosion. In 1952, Coward and Jones [6] gave a fast and simple method to get the possibility of explosion of mixed gas. Although this method is widely used, there is a restriction that the density of reactants should be known. Jian-Wei Cheng and Sheng-qiang Yang [7] improve the range, but the method needs the concentration of each component in the mixed gas.( see Figure 1). However, in the practical oil storage and production, the compositions and densities are influenced by many factors, such as, oil production purity, temperature and pressure. And in practice, the components in the mixture are unknown, neither the densities.

Figure 1: Coward explosive triangle aera A: impossible mixture B: explosive C1: not explosive(explosive if mixed more HC) C2: not explosive(explosive if mixed more air) D: non-explosive

With the development of artificial intelligence, nowadays the techniques of data mining are widely used into every field of chemical industry. Especially, the techniques like artificial neural network, genetic algorithm and fuzzy set theory have a lot of applications in equipment failure detection ([8] [9]). Since our goal is to accurately predict whether explosion of oil gas happens or not, and Support Vector Machine (SVM) and Logistic Regression (LR) are efficient to tackle the classification and prediction problems([10, 11]), we decide to use the two to make prediction of oil gas explosion. Artificial intelligence methods can improve accuracy of prediction and reduce man-made interference in which the selections of proper learning algorithms are very important. In recent years, SVM [10] is considered as an efficient learning algorithm about pattern classification [11], and the classification accuracy of SVM is better than other methods [12]. But the explicit expression of SVM is complex, so we can take another classification methodlogistic regression [13] into consideration. LR has a clear explicit expression, thus it can be 2

used to identify explosion intervals under different concentration of oxygen. The structure of the paper is as follows. In Section 2, we introduce the component of oil gas, resource of data. In Section 3, we introduce some basic knowledge about SVM and logistic regression. Some experiment results are performed in Section 4, the effect of penalty factor in SVM and comparison of two methods are also considered. Finally, we summarize the paper in Section 5. 2. Oil gas explosion In this section, we briefly give some preliminary knowledge about oilcan explosion, and introduce the explosion experiments, where the practical data are from a fire company. 2.1. Preliminary knowledge of oilcan explosion The three essential requirement of oil gas explosion are:(1) Oxygen; (2) Combustibles; (3) kindling. As static electricity has become a not easy-controlled factor, we tend to control the concentration of oxygen to prevent explosion of combustion. The limit of explosion is affected by a lot of factors, such as, temperature, oxygen content, inert medium, pressure, container, concentration of reactants. Reactants mainly contain HC compound, oxygen. As the explosion occurs only when the concentration of hydrocarbon (oil gas) and oxygen reaches some mixed ratio [5]. we establish a model to predict explosion with experimental data by data mining technique, based on the relationship between the concentration of hydrocarbon and oxygen. The prediction model is suitable for general oil gas without knowing densities. 2.2. Source of data The data are recorded in specific explosion experiments, including the explosion pressure and explosion situation. The experiments were conducted in special closed pipe at normal atmosphere pressure and room temperature, the device is showed in Figure 2.

Figure 2: Experiment Device

Oil gas is a mixture and the concentration of each component is unknown, we consider the total concentration of HC, the concentration of O2 and the concentration of premixed gas-mainly CO2 respectively. During each experiment, the concentration of O2 is controlled 3

by injecting CO2 , and the densities of reactants are detected at three different points and the explosion pressure are measured at five different points by sensors, the average concentration is taken as the concentration of reactants in one experiment, so is the maximum pressure. When the data are used to predict whether explosion occurs or not, we transform them into 0 or 1 in LR and −1 or 1 in SVM. In both of them, 1 corresponds to the explosion, others represent no explosion. The statistic data is showed in following tables. Table 1 lists one experiment result in which explosion occur and Table 2 lists another experiment that explosion does not occur. After taking several experiments, we get 58 groups of data. Finally we have 58 items and 5 variables. Since O2 /HC has a huge effect in reaction, so we also take it as a variable.

Meas. slowromancapi@ slowromancapii@ slowromancapiii@ Average

Table 1: An example of explosion data The data before ignite The data after ignite HC O2 CO CO2 HC O2 CO CO2 1.88 15.6 0.05 18.3 0.69 0.9 10.00 20.0 1.85 15.5 0.05 18.3 0.73 0.8 10.00 20.0 1.78 15.6 0.04 18.5 0.65 0.3 10.00 20.0 1.84 15.6 0.05 18.4 0.69 0.7 10.00 20.0

Pressure(%) P1 0.080 P2 0.080 P3 0.080 P4 0.080 P5 0.080

Meas. slowromancapi@ slowromancapii@ slowromancapiii@ Average

Table 2: An example of The data before ignite HC O2 CO CO2 1.81 15.3 0.60 13.2 1.82 15.3 0.62 13.3 1.81 15.4 0.62 13.4 1.81 15.3 0.61 13.3

Pressure(%) P1 0.000 P2 0.000 P3 0.000 P4 0.000 P5 0.000

unexplosion data The data after ignite HC O2 CO CO2 1.81 15.2 0.60 13.2 1.81 15.2 0.62 13.2 1.83 15.2 0.62 13.4 1.82 15.2 0.61 13.3

3. Implemental procedure of SVM and Logistic Regression In this section, we introduce the detailed procedure of explosion prediction using SVM and LR methods. 3.1. Data preprocessing Real world data is generally incomplete, noisy, and inconsistent. After identifying input and output variables, which are introduced in Section 2, data need to be preprocessed. In the data preprocessing, the noisy and incomplete data are removed. And inconsistencies are corrected in the data. For further considerations, different monitoring attributes have different scales. We need to normalize all attributes values into the same scale to avoid the influence of scales. All values of attributes are normalized to the interval [-1, 1] by using the Eq. (1). In this equation, suppose we have n samples, vij is the value of i-th attribute of j-th object and vij′ is the normalized value. max j = max {v1j , v2j , . . . , vnj }, min j = max {v1j , v2j , . . . , vnj }. 4

In summary, data preprocessing techniques can improve the quality of the data, so that improve the accuracy and efficiency of data mining process. vij′ =

2 max j + min j vij − max j − min j max j − min j

(1)

3.2. Feature selection In this subsection, we use some methods to select the key features. Some knowledge of chemical industry is used to select the key features. In the logistic regression, we use the generalized linear regression. The correlation between variables in data mining and whether the oil and gas will explode only relaying on the original variable, we expand the original data to quadratic and proportional items. However, there will be much more variables if we expand the original variables, so we need to reduce the dimension of the variables. Typically, there are two kinds of algorithms to reduce the feature space in classification. The first one is feature selection which is to select a subset of most representative feature from the original feature space. The second algorithm is feature extraction which is to transform the original feature space to a smaller one to reduce the dimension. Although feature extraction can reduce the dimension of feature space greatly compared with feature selection, the transformed smaller feature space cannot be explainable. In the test, we use the first method of feature selection; we use the all subset selection and Lasso ([14]) combined to the criterion of AIC, BIC. However, the experimental result is not good, and the difference between the original variable and variable selected by the criterion above is slight. Due to the obvious characteristics of the variable and the fact that the number of the variable is few, general methods of feature selection are not effective. In this paper, we use the knowledge of chemical industry and the role of the reactions to select the variable. In the reactions, CO is nearly 0 and it has no influence to the react. The role of CO2 is to control the concentration of O2 . Thus in the paper, we do not consider the effect of CO and CO2 . Then in the reactions, there are only two variables, HC and O2 . Because the ratio between HC and O2 has serious effect on the reaction pressure in the oil and gas reactions, we take the ratio between HC and O2 as a variable. In the end, we get the variables, HC, O2 and O2 /HC, denoted by x1 , x2 , x5 , respectively. 3.3. Learning algorithms In this subsection, we introduce the procedure of learning algorithms for prediction, SVM and LR. 3.3.1. Support vector machines As well known, SVM has been employed as the learning algorithm due to its superior classification ability. SVM is a supervised learning technique and it can be used for classification and regression. The main advantage of SVM include the use of kernel (no need to acknowledge the non-linear mapping function), the absence of local minima (quadratic problem), the sparseness of solution and the generalization capability obtained by optimizing the margin [15]. 5

Briefly speaking, SVM establishes a decision boundary between two classes by mapping the training data (through the kernel function) into a higher dimensional space, and then finding the maximal margin hyperplane within that space, which can be viewed as a classifier. Further introduction to SVM operations can be found in the following. Given n examples, S = {xi , yi }ni=1 , where xi represents the condition attributes, yi is the class label, and i is the number of examples. The decision hyperplane of SVM can be defined as (ω, b), where ω is a weight vector and b a bias. Let ω0 and b0 denote the optimal values of the weight vector and bias. Correspondingly, the optimal hyperplane can be written as ω0T ω + b0 = 0

(2)

To find the optimum values of ω and b, it is required to solve the following optimization problem. n P εi min 12 ω T ω + C ω,b,ε (3) i=1 T s.t. yi ω φ(xi ) ≥ 1 − εi , εi ≥ 0 Where ε is the slack variable, C is the user-specified penalty parameter of the error term(C > 0), and φ is the kernel function. To sum up, SVM can change the original non-linear separation problem into a linear separation case by mapping input vector onto a higher feature space. In addition, several popular kernel functions are listed as Eqs. (4)-(7). Linear kernel K(xi , xj ) = xi xj (4) Polynomial kernel of degree K(xi , xj ) = (γxi xj + r)g , γ > 0

(5)

K(xi , xj ) = exp(−γ k xi − xj k2 ), γ > 0

(6)

K(xi , xj ) = tanh(γxi xj + r), γ > 0

(7)

Radial basis function Sigmoid kernel Here, r, γ and g are kernel parameters and they are user-defined. According the work of Hsu et al. [16], RBF kernel function is selected in this study. Readers can find more details about SVM in [17, 18, 11]. 3.3.2. The penalty parameters of SVM When we use the model to predict the explosion, there are two categories of errors. The first one is the situation that the prediction result is 1 but the actual result is -1; the second one is the situation that the prediction result is -1 but the actual result is 1. SVM can set penalty upon the two types of error. Suppose ω1 and ω2 are the penalty parameters

6

of the two types of error, then the objective function of SVM that have different penalty parameters can be written as 1 T ω ω + ω1 2

X

ε j + ω2

{j:yi =1,ωxj +b=−1}

X

εm

(8)

{m:ym =−1,ωxm +b=1}

In 4.3, we will discuss the selection of penalty parameters in details. 3.3.3. Logistic Regression The method of general least square is not appropriate, we can use nonlinear function. LR is a classic nonlinear classifier algorithm. The advantage of LR is that it can predict the probability of the explosion of oil gas [13]. Given n examples, S = {xi , yi }ni=1 , yi ∈ {0, 1}, where xi represents the condition attributes, yi is the class label, and i is the number of examples. LR model is to get the probability function p(y = 1 | X = x) =

1 exp(x′ β) = ′ 1 + exp(x β) 1 + exp−g(x)

(9)

Where p is the probability of y = 1 while the example chooses x which is the conditional attribute, and g(x) = x′ β and β is the parameter of LR [13]. If p > 0.5, then y = 1; if p < 0.5, then y = 0. 3.3.4. Interval prediction using LR As the concentration of CO is almost 0, we omit its influence. Since the role of CO2 is to control the concentration of O2 , we consider O2 instead of CO2 . So in accordance with Eq. (9), we get a probability function p = p(HC, O2) for the concentration of HC and O2 . From the function, we can get the probability of explosion, and also the explosive range of the concentrations of oil gas when the concentration of oxygen is fixed relatively. 4. Computational results In this section, we give some experimental results. Table 3 gives a brief explanation of the data background, which includes size of data, number of variables and classification. There are totally 58 groups of data used to analysis. Among them, there are five input variables and one output variable which is binary variables-(explosion, not explosion). In order to establish the model of prediction, we select 80% of the data as training set, 20% of data as test set. Table 4 summarizes the variables in dataset. As stated in the above, we select x1 , x2 , and x5 as investigative variables. 4.1. Explosion intervals of LR According to above, we use logistic regression model to obtain the explosion intervals, which are illustrated in Figure 3 and Figure 4. The y-axis denotes the probability of explosion with the respective x under the given concentration of O2 , and Figure 4 shows the relative 7

Table 3: The characteristics of dataset for thermal for storage tank

No. of records 58 explosion:78% not explosion:22%

input attributes Target 5 Attributes 1 target attribute with 2 classes

Table 4: The summary of attributes

Type Measure Measure Ratio

description HC CO HC/O2

notation x1 x3 x5

Type Measure Measure Status

description notation O2 x2 CO2 x4 explosion or not y

probability of explosion with plot of g(x). The limit densities of explosion of HC are summarized in Table 5. Here we care whether the explosion occurs and the explosion intervals. So we just compare the intervals in horizontal axis. The result performs well, it means that the logistic regression can give an intuitive explanation.

Figure 3: probability

4.2. Comparisons between SVM and LR We use the popular RBF kernel in SVM, because of its good performance [19]. RBF kernel is a uniform framework which includes other kernels and eliminate troubles caused 8

Table 5: The limit densities of explosion of HC

Concentration of O2 15 16 18 20

the lower limit 1.0668 0.89729 0.76653 0.70066

the upper limit 1.5491 1.9645 2.5871 3.1448

Figure 4: relative probability

by huge data size. As it is well known, the selection of SVM parameter has much to do with prediction accuracy. In our model, C and λ in SVM can be adjusted. In this study, LIBSVM-3.1 [20] is used, which is one of the most popular SVM software. In order to compare the two methods, we take one time and multi-times experiments of vfold cross-validation (here, we take v=5). In one time experiment, by 5-fold cross-validation, we get 5 accuracies of prediction and the average accuracy. In multi-times experiments, we just consider the average accuracy of prediction in one 5-fold cross-validation. The results of one time 5-fold cross-validation are illustrated in Table 6. The figure of one time cross-validation is showed in Figure 5. (The 6-th point represents the average of the 5 points ahead.) After one-time cross-validation, we take ten cross-validation of SVM and LR and compare the average accuracy, which are showed in Table 7 and Figure 6. From the table 7 and figure 6, it can be seen that the prediction accuracy of SVM is better than logistic regression, and logistic regression is not very stable. The minimum accuracy of SVW and LR are 91% and 86% respectively. And the average accuracy of the two model 9

Table 6: The summary of one time 5-fold cross-validation using SVM and LR

Method Performance i-th one cross-validation 1 2 3 4 5 Mean/6 Std

SVM Accuracy (%) 85 100 100 92 90 93 6.54

LR Accuracy (%) 80 100 100 70 100 90 13.04

Figure 5: details in one 5-fold cross validation

are 91% and 87% respectively. From the point view of variance, the variance of accuracy in SVM model is 0.94, which is 1.77 in LR model, and in one time 5-fold cross-validation, the value are 6.54 and 13.04 respectively. According to both average and variance of accuracy, the prediction performance of SVM is better than that of logistic regression. 4.3. Set penalty parameters of SVM In this section, different penalty parameters setting on the two types of error are considered.

10

Table 7: The summary of 10 times 5-fold cross-validation using SVM and LR

Method Performance i-th cross-validation 1 2 3 4 5 6 7 8 9 10 Mean Std

SVM LR Accuracy (%) Accuracy (%) 92 91 93 90 91 89 91 88 90 87 90 86 91 87 91 86 91 86 90 86 91 87.6 0.94 1.77

Figure 6: multi-times of 5-fold cross validation

4.3.1. Different errors in oil gas explosion prediction In the procedure of prediction, it will arise two types of error. The first one is that the prediction result is 1 but the actual result is -1; the second one is that the prediction result is -1 but the actual result is 1. In our work, the first type of error means that the prediction result is not explosive but the actual result is explosive; the second type of error means on 11

the contrary. One can find that the first error is more serious and it may lead big dangerous accident. Hence, we prefer to reduce the first error in practical predictions. Although these two types of error are inevitable, we can change the distribution of the types of error by adjusting parameters in SVM. The effect of the two types of error always has a big difference. From our opinion, the first type of error is more disastrous. So we pay more attention to avoid the first type of error. According to SVM model, in Eq. 8, we set ω1 > ω2 , where ω1 is the penalty parameter of the first type of error and ω2 is that the second type of error. From above discussion, our goal is to maximize accuracy of prediction, and meanwhile minimize the first type error, which implies that the selection of the ratio of ω1 and ω2 are quite important. In our experiments, we try to control the distribution of the two types of error by adjusting the ratio of ω1 and ω2 . 4.3.2. Balance experiments of two types of error In this section, we choose the appropriate ω1 and ω2 to control the distribution of the two types of error. In our experiment, we construct the model when ω1 = ω2 and ω1 = γω2 , where γ ∈ [1, 60] denotes the loss ratio coefficient. In this model, we take radial basis function as the kernel function. The result of the distribution of the two types of errors and the whole error rate are illustrated in Figure 7 with the stepsize of γ is 5, i.e., γ = 5k, k = 1, . . . , 12. As the number of data in test set is fixed, the error rate can represent the number of errors.

Figure 7: Error Rate v.s. Weight ratio

12

In Figure 7, one can find that the first error rate is larger than the second with ω1 = ω2 , which is not the case what we expect. When γ > 0 increases, we can find that the first-type error rate is less than the second one. We can also find the intervals of γ corresponding the lowest first-type error as [5, 5], [8, 11], [15, 25] and [45, 60]. The we prefer to choose the loss ratio coefficient in these intervals such that the the whole error rate is as low as possible. In fact, we have choices of γ such that the whole error rate is 9% and the first-type error rate is only 3%. This result can control the distribution of the two types of errors and can achieve the goal of protecting people’s life although it may loss some profit for the second errors. 5. Discussion In this paper, we study the explosion prediction of oil gas, in which components are unknown. We mainly concern two problems, one is to predict whether the oil gas will explode, the other is to predict the explosion interval of the given concentration of mixed HC. The different statistic methods are introduced to analysis the real experimental data. Basis the results of statistic analysis, one can predict the explosion of the given certain percentage composition of oil gas and oxygen. Precisely, we mainly introduce the methods of SVM and LR in this paper. LR method can give an explicitly probability expression. The probability expression can be used for the tanks whose data of the components of oil gas are relatively stable and having no real-time updating database. SVM method can keep learning the data with updating. Comparing to LR, SVM can get higher accuracy. And SVM applies to the area of oil tanks with the data being collected real-time. As known that it is hard to detect the percentage composition of each component in oil gas, our models concern the total percentage composition of all oil gas to predict the explosion. Thus the methods are suitable for abroad applications. Before the explosion intervals estimating and the explosion predicting, we normalize the data and select variables for discussion. By using practical experimental data, we select the variables according to the purpose and background of the experiments. Using LR, one can obtain the explosion intervals of concentration. Considering the different levels of concentration of oxygen, we give the explosion intervals of concentration of oil gas respectively. From the computational results, we can find that the intervals of explosion are expanded to the two ends along with the increase of the concentration of oxygen. This situation just coincides with practical experience. With cross-validation in numerical experiments, the average prediction accuracy of LR can reach 87%, while the SVM gets more than 90%. These results show the applicability of SVM and LR for predicting the explosion of oil gas. In practical security protection with some special purposes, higher accuracy of prediction for some situations is needed. Generally, there are two kinds of error in prediction. One is that the no explosion is predicted but the data is in the range of explosion actually. Another is that the explosion is predicted for the new data but it is not true. We cannot tell which is more important of the different errors in theoretical analysis. However, in practical application, the second type of error may induce oil tank fire, which leads to disastrous 13

consequences. According to practical needs, we can restrict its occurrence as less as possible to meet the practical security requirements. Acknowledgement The authors would like to thank the Professor Sanguo Zhang for his useful advice and comments on this paper. We also thank Weitelog Fire Company for the data experiments. This work is supported by National Key Basic Research Program of China under Grant No.2011CB706901. References [1] D. Rasbash, Review of explosion and fire hazard of liquefied petroleum gas, Fire Safety Journal 2 (1980) 223 – 236. [2] S. Hφiset, B. H. Hjertager, T. Solberg, K. A. Malo, Statistical estimation of loads from gas explosions, Journal of Loss Prevention in the Process Industries 10 (1997) 271 – 283. [3] J. I.Chang, C.-C. Lin, A study of storage tank accidents, Journal of Loss Prevention in the Process Industries 19 (2006) 51–59. [4] CCPS, Guidelines for Chemical Process Quantitative Risk Analysis, Wiley-AIChE, second edition, 1999. [5] N. Wang, Oil gas and transport ship, Wuhan University of Technology Lecture, 2009. [6] H. Coward, G. Jones, Limits of flammability of gases and vapors, U.S. Bureau of Mines, Bulletin 503.US Government Printing Office (1952) 153. [7] J. Cheng, S. Yang, Improved coward explosive triangle for determining explosibility of mixture gas, Process Safety and Environmental Protection 89 (2011) 89–94. [8] A. Evsukoff, S. Gentil, Recurrent neuro-fuzzy system for fault detection and isolation in nuclear reactors, Advanced Engineering Informatics 19 (2005) 55–66. [9] M. C. Garcia, M. A. Sanz-Bobi, avier del Pico., Simap: Intelligent system for predictive maintenance application to the health condition monitoring of a windturbine gearbox, Computers in Industry 57 (2006) 552–568. [10] V. N.Vapnik, The Nature of Statistical Learning Theory, Springer, 1999. [11] Y. Liu, Y. Chen, Face recognition using total margin-based adaptive fuzzy support vector machines, IEEE Trans Neural Networks 18 (2007) 178–192. [12] K. Pload, S. Gunes, A. Arslan, A cascade learning system for classification of diabetes disease: generalized discriminant analysis and least square support vector machine, Expert Systems with Applications 34 (2008) 482–487. [13] J. Wang, Z. Guo, Logistic Regression Model–Method and applications, Higher Education Press, 2001. [14] R. Tibshirani, Regression shrinkage and selection via the lasso, J. Royal. Statist. Soc B. 58 (1996) 267–288. [15] A. Cerqueira, D. Ferreria, M. Ribeiro, C. Duque, Power quality events recognition using a svm-based method, Electric Power Systems Research 78 (2008) 1546–1552. [16] C. Hsu, C. Chang, C. Lin, A practical guide to support vector classification (2006). [17] B. Asl, S. Setarehdan, M. Mohebbi, Support vector machine-based arrhythmia classification using reduced features of heart rate variability signal, Artificial intelligence in medicine (2008). [18] Y. Liu, H. Huang, C. Weng, Recognition of electromyographic signals using cascaded kernel learning machine, IEEE/ASME Transactions on Mechatronics 12 (2007) 253–264. [19] K. Coussment, K. Poel, Churn prediction in subscription service: an application of support vector machines while comparing two parameter selection techniques, Expert Systems with Applications 34 (2008) 331–327. [20] C. Chang, C. Lin, Libsvm: a library for support vector machines (2001).

14