Copula selection for graphical models in ... - Semantic Scholar

2 downloads 0 Views 854KB Size Report
Abstract This paper presents the use of graphical models and copula ... graphical models for modeling the dependencies among variables provides some the-.
Comput Stat DOI 10.1007/s00180-013-0457-y ORIGINAL PAPER

Copula selection for graphical models in continuous Estimation of Distribution Algorithms Rogelio Salinas-Gutiérrez · Arturo Hernández-Aguirre · Enrique R. Villa-Diharce

Received: 14 March 2011 / Accepted: 1 October 2013 © Springer-Verlag Berlin Heidelberg 2013

Abstract This paper presents the use of graphical models and copula functions in Estimation of Distribution Algorithms (EDAs) for solving multivariate optimization problems. It is shown in this work how the incorporation of copula functions and graphical models for modeling the dependencies among variables provides some theoretical advantages over traditional EDAs. By means of copula functions and two well known graphical models, this paper presents a novel approach for defining new EDAs. Either dependence is modeled by a copula function chosen from a predefined set of six functions that aim to cover a wide range of inter-relations. It is also shown how the use of mutual information in the learning of graphical models implies a natural way of employing copula entropies. The experimental results on separable and non-separable functions show that the two new EDAs, which adopt copula functions to model dependencies, perform better than their original version with Gaussian variables. Keywords Multivariate optimization problems · Copula functions · Copula entropies · Likelihood function 1 Introduction Estimation of Distribution Algorithms (EDAs) are a recent paradigm in Evolutionary Computation (EC) for solving optimization problems through the incorporation of R. Salinas-Gutiérrez (B) Universidad Autónoma de Aguascalientes, Aguascalientes, México e-mail: [email protected]; [email protected] A. Hernández-Aguirre · E. R. Villa-Diharce Center for Research in Mathematics, Guanajuato, México e-mail: [email protected] E. R. Villa-Diharce e-mail: [email protected]

123

R. Salinas-Gutiérrez et al.

probabilistic models. In particular, the selection of the multivariate normal distribution has been a common choice in EDAs to model continuous decision variables in optimization problems. However, this selection may not be adequate because marginal distributions must be assumed as normal distributions along with the assumption of linear dependencies among variables. These restrictive assumptions can be avoided by means of the proper use of copula functions. This work shows how the incorporation of the copula theory and its graphical representation can increase the capacity of an EDA to model the most important dependencies. Similar to Genetic Algorithms (GAs), EDAs are based on populations. However, this new class of evolutionary optimization techniques do not use genetic operators such as crossover and mutation for generating new individuals. Instead, in EDAs, the new individuals are sampled from a probability distribution. Therefore, the goal in EDAs is to model the dependencies of the best individuals and transfer them into the next population. A pseudocode for EDAs is shown in Algorithm 1. Algorithm 1 Pseudocode for EDAs 1: Assign t ←− 0. Generate the initial population P0 with N individuals at random. 2: Evaluate population Pt using a cost function. 3: Select a collection of M individuals St , with M < N and according to the selection method, from Pt . 4: Estimate a probabilistic model Mt from St . 5: Generate the new population by sampling from Mt . assign t ←− t + 1. 6: If stopping criterion is not reached go to step 2.

According to the pseudocode presented in Algorithm 1, an individual is equivalent to a multivariate observation whereas the cost function is related to the optimization problem. The stopping rule can be a maximum number of evaluations from the cost function or some convergence criterion. The main advantage of using probabilistic distributions in evolutionary algorithms is that the interrelations among the variables of a population are explicitly modeled. Thus, according to step 1 of Algorithm 1, the estimation of a probabilistic model is an important procedure in EDAs. Therefore, ever since EDAs were introduced in the field of EC by Mühlenbein and Paaß (1996), many researchers have been interested in proposing and enhancing new probabilistic models. Nowadays there are several EDAs for optimization problems in discrete and continuous domains. The EDAs can be classified as univariate, bivariate or multivariate according to the complexity of their probabilistic model used to learn the interactions between the variables. The univariate EDAs consider all the variables independent, for instance, the Univariate Marginal Distribution Algorithm (UMDA) (Mühlenbein 1998; Larrañaga and Lozano 2002), the Population Based Incremental Learning (PBIL) (Baluja 1994), and the compact Genetic Algorithm (cGA) (Harik et al. 1998). The bivariate EDAs take into account dependencies between pairs of variables and a few examples are the Bivariate Marginal Distribution Algorithm (BMDA) (Pelikan and Mühlenbein 1999), Mutual Information Maximizing Input Clustering (MIMIC) (De Bonet et al. 1997; Larrañaga and Lozano 2002), and Dependency-Trees (Baluja and

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

Davies 1997). Many univariate and bivariate discrete EDAs have been extended to continuous domains by using Gaussian probabilistic models. For multiple dependencies in discrete domain the EDAs have used probabilistic models such as the Polytree Approximation of Distribution Algorithm (PADA) (Soto et al. 1999), Estimation of Bayesian Networks Algorithm (EBNA) (Larrañaga and Lozano 2002) and Bayesian Optimization Algorithm (BOA) (Pelikan et al. 1999). For real-valued (continuous) multivariate variables the EDAs have used mostly multivariate Gaussian distributions and some examples are the Estimation on Multivariate Normal Algorithm (EMNA) (Larrañaga and Lozano 2002) and Estimation of Gaussian Network Algorithm (EGNA) (Larrañaga and Lozano 2002). The EDA AMaLGaM (Bosman et al. 2008) and the algorithm CMA-ES (Igel et al. 2006) are also based on the multivariate Gaussian distribution. Both algorithms modify the estimated covariance matrix to improve the convergence rate towards the optimum. Currently they are the state of the art in real-valued optimization. Nowadays there has been very few works related to the study and incorporation of copula functions into EDAs. Moreover, given that the investigation into copula functions and graphical models is also a young research area, the use of copula functions and graphical models in EDAs is almost null. This work presents two novel EDAs for solving optimization problems based on graphical models which are built with different copula functions. For the interested reader in copula theory, this paper provides a potential application of copula functions for solving optimization problems by means of EDAs. However, it is important to note that the main purpose and aim of this work is precisely the solution of optimization problems with EDAs based on the incorporation of copula functions. Furthermore, EDAs based on graphical models may easily adopt copula functions in various ways that is harder to achieve by other algorithms which are strictly oriented to model relationships as linear dependencies. In this paper, a chain and a tree are used as graphical models and the related variables are modeled by the most appropriate copula function. The goals of this work are (1) to use copula functions for modeling the most important dependencies or associations between variables, (2) to take advantage of the capacity of copula functions to isolate dependency structure from the marginal behaviour of individual variables, and (3) to study experimentally the performance of the proposed EDAs. The proposed EDAs use a procedure based on the loglikelihood function to determine which copula function will be chosen to model the relationship between variables. Related works have considered EDAs based on Gaussian copula function (Wang and Zeng 2010) and EDAs based on Archimedean copula functions with a fixed dependence parameter (Wang et al. 2010a,b; Cuesta-Infante et al. 2010; Wang and Zeng 2010). Unlike the previous papers that use Archimedean copula functions, a way of estimating the copula parameter is presented in (Gao 2009). These works use copula functions to model all pairwise relationships among variables and do not employ graphical models. On the other hand the works (Salinas-Gutiérrez et al. 2009, 2010) use Frank and Gaussian copula entropies in order to get a graphical model and establish the most important dependencies between variables. To the best of our knowledge, the works (Gao 2009; Salinas-Gutiérrez et al. 2009, 2010) are the only ones that employ the maximum likelihood method for estimating the copula parameters.

123

R. Salinas-Gutiérrez et al.

The structure of the paper is the following: Sect. 2 is a brief introduction to bivariate copula functions, Sect. 3 describes the implementation of copula functions in the chain and tree graphical models, Sect. 4 explains the selection procedure, Sect. 5 presents the experimental setting to solve a test suite for global optimization, and Sect. 6 resumes the conclusions. 2 Copula theory The copula functions are suitable tools in statistics for modeling dependencies, not necessarily linear dependence, in several random variables. The copula theory was introduced by Sklar (1959) to separate the effect of dependence from the effect of marginal distributions in a joint distribution. Although copula functions can model linear and nonlinear dependencies, they have been barely used in computer science applications where nonlinear dependencies are common and need to be represented. Definition 1 A copula function is a joint distribution function of standard uniform random variables. That is, C(u 1 , . . . , u d ) = Pr [U1 ≤ u 1 , . . . , Ud ≤ u d ] , where Ui ∼ U (0, 1) for i = 1, . . . , d. As a consequence of Definition 1, the copula density can be calculated as: c(u 1 , . . . , u d ) =

∂ d C(u 1 , . . . , u d ) . ∂u 1 · · · ∂u d

(1)

The interested reader in a more formal definition of copula function is referred to Joe (1997), Nelsen (2006), Trivedi and Zimmer (2007). The following result, known as Sklar’s theorem, gives the relevance and practical utility to copula functions. Theorem 1 (Sklar) Let F be a d-dimensional distribution function with marginals d F1 , F2 , . . . , Fd , then there exists a copula C such that for all x in R , F(x1 , x2 , . . . , xd ) = C (F1 (x1 ), F2 (x2 ), . . . , Fd (xd )) , where R denotes the extended real line [−∞, ∞]. If F1 (x1 ), F2 (x2 ), . . . , Fd (xd ) are all continuous, then C is unique. Otherwise, C is uniquely determined on Ran(F1 ) × Ran(F2 ) × . . . × Ran(Fd ), where Ran stands for the range. According to Theorem 1 and using the chain rule for differentiating composite functions along with the Eq. (1), any d-dimensional density f can be represented as f (x1 , . . . , xd ) = c (F1 (x1 ), . . . , Fd (xd )) ·

d  i=1

123

f i (xi ),

(2)

Copula selection for graphical models in continuous Estimation of Distribution Algorithms Table 1 Bivariate copula distribution functions Copula

Description

Clayton

uv ; θ ∈ [−1, 1) 1 − θ (1 − u)(1 − v)     3θ − 2 2 1 2 τ= − 1− ln(1 − θ ) 3θ 3 θ   C(u, v) = max (u −θ + v −θ − 1)−1/θ , 0 ; θ ∈ [−1, ∞)\{0}

Farlie–Gumbel–Morgenstern

θ θ +2 C(u, v) = uv (1 + θ (1 − u)(1 − v)) ; θ ∈ [−1, 1]

Ali-Mikhail-Haq

C(u, v) =

τ=

τ=

Frank

Gaussian

Gumbel

2 θ 9

  (e−θu − 1)(e−θv − 1) 1 ; θ ∈ (−∞, ∞)\{0} C(u, v) = − ln 1 + θ e−θ − 1 4 t 1 θ τ =1− 1− dt 0 t θ θ e −1 1  −1

Φ −1 (u) Φ −1 (v) e− 2 t Σ t dt1 dt2 ; θ ∈ (−1, 1) C(u, v) = −∞ −∞ 2π |Σ|1/2

where Σ is a correlation matrix with Σ12 = θ 2 τ = sin−1 (θ ) π

C(u, v) = exp −(u˜ θ + v˜ θ )1/θ ; θ ∈ [1, ∞) where u˜ = −ln(u) and v˜ = −ln(v) 1 τ =1− θ

where c is the density of the copula C, and f i (xi ) is the marginal density of variable xi . Equation (2) shows that the dependence structure is modeled by the copula function. This expression separates any joint density function into the product of copula density and marginal densities. This contrasts with the usual way to model multivariate distributions, which suffers from the restriction that the marginal distributions are usually of the same type. The separation between marginal distributions and a dependence structure explains the modeling flexibility given by copula functions. In this paper we use two-dimensional parametric copula functions for modeling the dependence structure of random variables associated by a joint distribution function. Table 1 shows the distribution functions of the bivariate copula functions used in this work as well as the appropriate values for the dependence parameter θ . These copula functions have been chosen because they cover a wide range of dependencies. Six copula functions are available to choose from Table 1. These copula functions have appeared frequently in the literature and have been used in a broad range of applications. For instance, the Ali-Mikhail-Haq, Clayton, Farlie–Gumbel–Morgenstern, Frank and Gaussian copula functions can model negative and positive dependences

123

R. Salinas-Gutiérrez et al.

between the marginals. One exception is the Gumbel copula which does not model negative dependence. The Ali-Mikhail-Haq and Farlie–Gumbel–Morgenstern copula functions are adequate for marginals with modest dependence. When dependence is strong between extremes values, the Clayton and Gumbel copula functions can model left and right tail association respectively. The Frank copula is appropriate for data that exhibit weak dependence between extreme values and strong dependence between centered values, while Gaussian copula is adequate for data that exhibit weak dependence between centered values and strong dependence between extreme values. For parametric bivariate copula functions, the dependence parameter is related to Kendall’s tau through the following expression (see Nelsen 2006) 1 1 C(u, v; θ )dC(u, v; θ ) − 1.

τ (X, Y ) = 4 0

(3)

0

The dependence parameter θ of a bivariate copula function can be estimated using the Maximum Likelihood method (ML) and the estimation of Kendall’s tau. To do so, the one-dimensional log-likelihood function n    n = ln c(u i , vi ; θ ),  θ ; {(u i , vi )}i=1

(4)

i=1

is maximized using as starting point the nonparametric estimation of Kendall’s tau. Assuming the marginal parameters are known, the pseudo copula observations n {(u i , vi )}i=1 in Eq. (4) are obtained by using the marginal distribution functions of variables X and Y . Tables 1 and 2 show the formulas for Kendall’s tau and the copula densities, respectively. It has been shown in Weiß (2011) that the ML estimator has better properties than other estimators. For estimating all the parameters of a general joint probabilistic model, such as Eq. (2), the Full Maximum Likelihood method (FML) can be employed. The FML method simultaneously estimates the marginal parameters and the copula parameters by using the ML method. However, the FML method could be very computationally intensive for the estimation task. For this reason, the FML method can be substituted by a two step strategy in which the marginal distributions are estimated in the first step and the dependence parameters are estimated in the second step. Given that the estimation of the copula parameters requires assumptions on the marginal distributions, two general approaches (Trivedi and Zimmer 2007) can be considered for the marginal distributions in the two step strategy: (1) parametric distributions, and (2) nonparametric distributions. A fully parametric method based on the two step strategy is known as the Inference Function for Margins method (IFM), whereas the Canonical Maximum Likelihood method (CML1 ) uses nonparametric distributions, such as the empirical distribution, for calculating the pseudo copula observations (Cherubini et al. 2004). As shown in Kim et al. (2007), the CML method is conceptually 1 In the paper of Kim et al. (2007), the CML method is also named as Pseudo Maximum Likelihood (PML).

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms Table 2 Bivariate copula densities and conditional distributions Copula

Description

Ali-Mikhail-Haq

c(u, v) =

Clayton

v − θ v(1 − v) ∂C = ∂u (1 − θ (1 − u)(1 − v))2

−2−1/θ c(u, v) = (1 + θ ) (uv)−θ−1 u −θ + v −θ − 1

1 + θ (u + v + uv − 2) − θ 2 (u + v − uv − 1) (1 − θ (1 − u)(1 − v))3

− 1 −1 ∂C = u −θ−1 u −θ + v −θ − 1 θ ∂u Farlie–Gumbel–Morgenstern

c(u, v) = 1 + θ (1 − 2u)(1 − 2v) ∂C = v 2 (θ (2u − 1)) + v (1 + θ (1 − 2u)) ∂u

Frank

Gaussian

Gumbel

c(u, v) = 

−θ (e−θ − 1)e−θ(u+v)

2 (e−θu − 1)(e−θv − 1) + (e−θ − 1)

∂C e−θu (e−θv − 1) = −θu ∂u (e − 1)(e−θv −1) + (e−θ − 1) 

1/2 (x 2 + y 2 − 2θ x y) (x 2 + y 2 ) exp − c(u, v) = 1 − θ 2 + 2 2(1 − θ 2 ) where x = Φ −1 (u) and y = Φ −1 (v)

C(u, v) ˜ θ−1 (u˜ v) θ θ 1/θ + θ − 1 c(u, v) =  2−1/θ (u˜ + v˜ ) uv u˜ θ + v˜ θ  θ−1 ∂C C(u, v) lnu = ∂u lnC(u, v) u where u˜ = −ln(u) and v˜ = −ln(v)

almost the same as the IFM and it is robust against misspecification of the marginal distributions. In this work, optimization problems with continuous variables are considered. Thus, a semiparametric variant of the CML method presented in Trivedi and Zimmer (2007) is used in this work for (1) estimating the dependence parameter, and (2) generating new individuals with continuous attributes. In this variant, a nonparametric kernel density is used to estimate the univariate marginal densities. The steps of the semiparametric method used in this paper are shown in Algorithm 2. In order to appreciate the kind of dependence structure that copula functions can model, Fig. 1 shows a scatter plot with data drawn from a joint distribution with Gaussian marginals and structure dependence modeled by four different copula functions. All data sets have associated the same marginal distributions and the same value of Kendall’s tau. It can be appreciated that Fig. 1c shows a well known elliptical dependence structure. This is because the Gaussian copula is by definition the copula associated to the

123

R. Salinas-Gutiérrez et al.

Algorithm 2 Pseudocode for estimating the dependence parameter of a bivariate copula function with the semiparametric method 1: Use a Gaussian kernel density with a bandwidth parameter η for estimating the marginal density function f X of variable X . 2: Use a Gaussian kernel density with a bandwidth parameter ξ for estimating the marginal density function f Y of variable Y . n 3: Calculate the pseudo copula observations {(u i , vi )}i=1 using the marginal distribution functions of variables X and Y , u = FX (x; η), v = FY (y; ξ ). n 4: Estimate the copula parameter θ by using pseudo observations {(u i , vi )}i=1 and maximizing the loglikelihood function given by Equation (4).

(b)

−4

−4

−2

−2

0

0

2

2

4

4

(a)

−4

−2

0

2

4

−2

0

2

4

−4

−2

0

2

4

(d)

−4

−4

−2

−2

0

0

2

2

4

4

(c)

−4

−4

−2

0

2

4

Fig. 1 Four samples of 1,000 points that have been drawn from a a Clayton copula with θ = 4.667, b Frank copula with θ = 11.412, c Gaussian copula with θ = 0.891, and d Gumbel copula with θ = 3.333. All data sets have the same value of 0.7 for Kendall’s tau and use standard Gaussians as marginals

joint standard Gaussian distribution. When using Gaussian distributions as marginal distributions and the Gaussian copula as the dependence function, the joint distribution is a joint Gaussian distribution. Other kinds of dependence between Gaussian variables are possible by using different copula functions as shown in Fig. 1a–c.

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

For sampling from bivariate densities, the Algorithm 3 gives the steps of simulating variates. This Algorithm was used for getting the samples of Fig. 1. The conditional distribution functions for sampling from the bivariate copula functions are shown in Table 2. Algorithm 3 Pseudocode for generating a bivariate population with an associated copula function 1: Draw two independent random variables (u, t) from an uniform distribution U (0, 1). ∂C 2: Solve t = C V (v|u; θ ) for v, where C V (v|u; θ ) = is the conditional distribution of V given U . ∂u The pair (u, v) is a simulation from the bivariate copula function C(u, v; θ ). 3: Calculate X and Y using inverses of marginal distribution functions, x = FX−1 (u; η) and y = FY−1 (v; ξ ). Accept the pair (x, y).

It is also possible to use the conditional distribution Cv (v|u; θ ) for getting samples from a Gaussian copula. However, a well established method in copula theory for sampling from a Gaussian copula is presented in Algorithm 4. Algorithm 4 Pseudocode for generating a simulation from a Gaussian copula with parameter θ 1: Draw two independent random variables (w, t) from a standard Gaussian distribution N (0, 1).  2: Set z = w · θ + t 1 − θ 2 . 3: Calculate u = Φ(w) and v = Φ(z), where Φ is the cumulative standard Gaussian distribution function. Accept the pair (u, v).

3 Two graphical models based on bivariate copula functions Although copula functions can model dependencies among variables, sometimes it is not clear what multivariate copula function must be chosen. However, by means of graphical models (Koller and Friedman 2009) it is possible to model the most important dependencies or associations between variables. This is the case for a copula function, because a multivariate copula function is also a probabilistic model. The idea of using graphical models for factorizing a multivariate copula function was presented as Vines in the work of Bedford and Cooke (2002). Vines provide a specific factorization of a high-dimensional copula into bivariate copula functions. Further, vines have been used recently for modeling financial and economic data (Aas et al. 2009; Kurowicka and Cooke 2006). Although a truncated vine has been used as probabilistic model for defining a new EDA in Salinas-Gutiérrez et al. (2010), in this work we propose to incorporate copula functions into other graphical structures: a chain and a tree graphical models. These two graphical models are well known in EDAs and have not been used before for modeling multivariate distributions with copula functions. These models are based on pairwise conditional distributions and they are illustrated in Fig. 2.

123

R. Salinas-Gutiérrez et al.

Xα1

Xα3

Xα2

Xα4

Xα5

f (x) = f xα1 f xα2 |xα1 f xα3 |xα2 f xα4 |xα3 f xα5 |xα4

(a) Xα1

Xα3

Xα2

Xα5

Xα4

f (x) = f xα1 f xα2 |xα1 f xα3 |xα1 f xα4 |xα3 f xα5 |xα3

(b) Fig. 2 Two joint distributions over five variables represented by a a chain graphical model and b a tree graphical model

3.1 Chain model A chain graphical model for a d-dimensional continuous random vector X represents a probabilistic model with the following density:



f chain (x) = f xα1

d 

  f xαi |xα(i−1) ,

(5)

i=2

where α = (α1 , . . . , αd ) is a permutation of the integers between 1 and d. Figure 2a shows an example of a chain graphical model. Notice that a permutation could not be unique, in the sense that different permutations could give the same density values in (5). In practice the permutation α is unknown and the chain graphical model must be learnt from data. A way of choosing the permutation α is based on the Kullback–Leibler divergence. This divergence is an information measure between two distributions. It is always non-negative for any two distributions, and is zero if and only if the distributions are identical. Hence, the Kullback–Leibler divergence can be interpreted as a measure of the dissimilarity of two distributions. Then, the goal is to choose a permutation α that minimizes the Kullback–Leibler divergence between the true distribution f (x) of the data set and the distribution associated to a chain model, f chain (x). In this case the Kullback–Leibler divergence, D K L , of a continuous random vector X with joint densities f and f chain is given by:

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

D K L ( f || f chain ) = E f log

f (x) f chain (x)



= −H (X) + H (X α1 ) +

d 

  H X αi |X α(i−1) .

(6)

i=2

The first term in the divergence (6), H (X), is the entropy of the joint distribution f (x) and does not depend on the permutation α. The second term is the entropy of a marginal distribution. The third term represents the sum of conditional entropies. The marginal and conditional entropies are taken into account for minimizing the Kullback–Leibler divergence. However, it can be noticed that there is not a direct solution without considering all possible permutations. For this reason it is necessary to employ a greedy algorithm in order to find a sub-optimal permutation α. In the EDA literature, an algorithm that uses a chain graphical model is the MIMIC. This algorithm proposes a way to find a permutation in the following way: (1) set as root the variable with the lowest marginal entropy, and (2) choose the variable whose conditional entropy with respect to the previous variable is the lowest. The chain is built by repeating the last step with the rest of the variables. The MIMIC algorithm for continuous variables uses bivariate Gaussian distributions. This assumption gives a direct way of calculating marginal and conditional entropies. However, for some populations, the Gaussian assumption can not be the most adequate distribution for modeling them. In this paper we employ an equivalent expression for minimizing the Kullback– Leibler divergence (6). It is well known in information theory (Cover and Thomas 1991) that the conditional entropy between variables S and T is related to its mutual information I (S, T ) as follows: H (T |S) = H (T ) − I (S, T ), where

I (S, T ) = E f (s,t)

(7)

f (s, t) log . f (s) · f (t)

Substituting conditional entropies in Eq. (6) by the relation given by Eq. (7), the Kullback–Leibler divergence can be written as: D K L ( f || f chain ) = −H (X) +

d  k=1

d    H (X k ) − I X αi , X α(i−1) .

(8)

i=2

According to Eq. (8), minimizing the Kullback–Leibler divergence is equivalent to maximizing the total sum: Jchain (X) =

d    I X αi , X α(i−1) .

(9)

i=2

123

R. Salinas-Gutiérrez et al.

A sub-optimal permutation α can be found by taking into account just the total sum of pairwise mutual information. As we shall show later, the entropy of the copula function gives a natural way for computing the Eq. (9). 3.2 Tree model The joint density for a tree graphical model is given by: d     f tree (x) = f xβ1 f xβi |xβ p(i) ,

(10)

i=2

where β = (β1 , . . . , βd ) is a permutation of integers 1, . . . , d, and p(i) maps numbers 1 < i ≤ d to integers 1 ≤ p(i) < i. Each variable in Eq. (10) is conditioned upon at most one parent. It can be noticed that the chain graphical model is a special case of the tree graphical model when p(i) = i − 1. Figure 2b shows an example of a tree graphical model. In a similar way to the learning of a chain graphical model, a tree model can be learned by minimizing the Kullback–Leibler divergence between the true density function f (x) and the proposed density function, f tree (x): D K L ( f || f tree ) = −H (X) +

d 

H (X k ) −

k=1

d    I X βi , X β p(i) .

(11)

i=2

It is also possible to derive the expression (11) by using the Markov equivalence (Koller and Friedman 2009) between the directed version of a tree and its undirected counterpart: f tree (x) =



f i j (xi , x j )/

(i, j)∈T

d 

f i (xi )deg(i)−1 ,

(12)

i=1

where the degree, deg(i), is the number of adjacent nodes to node i. As can be seen, the number of bivariate densities in Eq. (12) is the same number of branches on the tree (10). These bivariate terms are taken into account for calculating the sum of mutual information in Eq. (11). The first two terms in Eq. (11) do not depend on the tree graphical structure. These two terms are related to the Kullback–Leibler divergence between the true density function and the independent model, i.e., the product of marginal densities. The third term is the sum of all mutual information associated to the d − 1 tree branches. Therefore, minimizing the Kullback–Leibler divergence is equivalent to maximizing the total sum: Jtree (X) =

d    I X βi , X β p(i) . i=2

123

(13)

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

The optimization problem (13) was presented in Chow and Liu (1968) as the total branch weight and it has been used for the learning of dependence trees. The optimal tree is the one that produces the highest total sum of mutual information. Since the amount of mutual information for every pair of nodes is independent of the rest, a greedy strategy such as the Kruskal’s algorithm (Cormen et al. 2009) can be used for maximizing the summation in Eq. (13). Although originally Kruskal’s algorithm was designed for finding a minimum spanning tree for a connected weighted graph, in an equivalent way, this algorithm enable us to build a maximum-weight dependence tree.

3.3 Copula entropy and mutual information By using copula functions, see Eq. (2), the joint density represented by a chain model can be written as d     f xαi |xα(i−1) f chain (x) = f xα1 i=2

      d  f xαi · f xα(i−1) · c u αi , u α(i−1)   = f xα1 f xα(i−1) i=2 

=

d  k=1

f (xk )

d    c u αi , u α(i−1) .

(14)

i=2

The above result is very illustrative. From Sklar’s theorem we know that there is an unique copula function which models the dependence structure. The Eq. (14) states that the copula density associated to a chain graphical model, cchain (u), can be decompose into the product of bivariate copula functions:

cchain (u) =

d    c u αi , u α(i−1) .

(15)

i=2

This is an interesting and practical consequence, because any multivariate copula function associated to a chain model can be built by employing the adequate bivariate copula functions. Moreover, the copula function associated to the chain model can be also graphically represented by a chain. To see that,

cchain (u) =

d    c u αi , u α(i−1) i=2

=

d    c u αi |u α(i−1) .

(16)

i=2

123

R. Salinas-Gutiérrez et al.

The previous calculations used the fact that the marginal distributions of a copula function are uniform distributions. It can be noted that Eq. (16) is similar to Eq. (5). For the case of a tree model it is not difficult to see the following: f tree (x) = ctree (u)

d 

f (xk ) ,

(17)

k=1

where d    ctree (u) = c u βi , u β p(i) .

(18)

i=2

The result in Eq. 18 states that the copula function associated to a joint distribution with a tree graphical model is also represented by a tree. Next, we discuss how the learning of a chain model or a tree model for the random vector X is related to the entropy of a copula function. We have seen that picking the sub optimal permutation for a chain model and the optimal permutation for a tree model depends on the total sum of pairwise mutual information, Eqs. (9) and (13). In this paper we propose the use of the following relationship, presented in Davy and Doucet (2003), between the mutual information of variables (S, T ) and the entropy of their associated copula function: I (S, T ) = −H (U, V ),

(19)

where copula variables (U, V ) are related to variables (S, T ) by their marginal distribution function, i.e., u = FS (s) and v = FT (t). Therefore, the Eq. (9) can be written as: Jchain (X) = −

d 

  H Uαi , Uα(i−1) .

(20)

i=2

A consequence of Eq. (20) is that a sub optimal permutation α for a chain model can be learned by using the sum of bivariate copula entropies instead of the sum of pairwise mutual information. This can be taken into account for proposing a new greedy algorithm for the MIMIC with copula functions. For instance, (1) choosing the two variables with the lowest copula entropy and set them the first elements in the chain, and (2) new chain links can be added to the chain by selecting variables with the lowest copula entropy according to the variables in the ends of the chain. This is shown in Algorithm 5. Algorithm 5 Greedy algorithm to pick a permutation α in a chain model () is an estimation of the copula entropy (U j , Uk ), where H 1: Choose (αm , αm+1 ) = arg min j=k H between the variables u j = FX j (x j ) and u k = FX k (xk ). 2: Choose variables with the lowest copula entropy with respect to any of the ends of the chain. The constraint is to avoid a circular chain. 3: The order of the chain defines permutation α.

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

For a tree graphical model is also possible to use the sum of pairwise mutual information for learning the optimal permutation β. This can be done by also using the Kruskal’s algorithm because the pairwise mutual information are replaced by the estimated copula entropies. The relationship between the entropy of a bivariate copula function and the marginal mutual information of two variables, Eq. (19), has both theoretical and practical importance: (1) the mutual information is given by the copula function regardless the marginal distributions, and (2) the estimation of the copula entropy can be more accurate than the estimation of mutual information because the copula domain is always bounded and standardized. In this work the mutual information is estimated by using a Monte Carlo simulation. Algorithm 6 illustrates the sampling procedure for estimating the entropy of a bivariate copula function. Algorithm 6 Monte Carlo method for estimating the copula entropy m from the copula distribution c(u, v) with dependence parameter 1: Simulate a random sample {(u i , vi )}i=1 θ. 2: Calculate m  (U, V ) = − 1 log c(u i , vi ; θ ) H m i=1

(U, V ) is a estimation of the copula entropy H (U, V ). The quantity H

4 Proposed EDAs In this paper it has been explained how to incorporate copula functions for modeling bivariate dependencies in tree branches and chain links. However, in order to gain modeling flexibility, this work also presents a procedure for selecting copula functions in EDAs. 4.1 The selection of a bivariate copula function Two well known tools in statistics for model selection are the Akaike information criterion (AIC) and the Bayesian information criterion (BIC). These criteria employ the maximized value of the likelihood function (Lmax ) for the estimated model, the number of parameters to be estimated (k) and the sample size (n): AIC = −2 ln(Lmax ) + 2k BIC = −2 ln(Lmax ) + k ln(n).

(21) (22)

For a particular chain link or tree branch, the selection of a copula function from Table 1 can be done just by means of any of the criteria in Eqs. (21) and (22). However, in each generation of an EDA, the selected population has a fixed size. Moreover, the bivariate copula functions from Table 1 have one parameter. Hence, it can be noted that

123

R. Salinas-Gutiérrez et al. Fig. 3 A tree with its associated copula functions over each branch

X1

Clayton

AMH X3

X2

Gaussian X4

Frank X5

for any of the criteria AIC and BIC, the selection of a copula function just depends on the maximized value of the likelihood function, Lmax . For that reason in this work2 we propose a copula selection procedure based on the highest value of the log-likelihood function [Eq. (4)]. The result of using copula selection in a tree is illustrated in Fig. 3. It can be noticed that, for all pairwise variables, their edges are considered for inclusion in the graphical model only if the corresponding copula entropies are negative. Otherwise, as illustrated in Abreu et al. (2010), the learnt model can be a forest.

4.2 Description of the proposed EDAs As already described in Sect. 3, it is possible to employ copula functions for representing a chain model or a tree graphical model. Moreover, as illustrated in Fig. 3, it is not necessary that all chain links or tree branches are modeled by the same family of copula functions. The main aspects, such as the estimation of the probabilistic model and the generation of the new population, are shown in Algorithm 7.

Algorithm 7 Pseudocode for estimating the probabilistic model and generating new population 1: for i = 1 to d do 2: Estimate the marginal density f i (xi ; ηi ) of variable X i 3: Calculate u i using the marginal distribution function Fi , u i = Fi (xi ; ηi ) 4: end for d(d − 1) bivariate copula parameters by using 5: For each copula function, estimate the corresponding 2 variables Ui and the maximum likelihood method d(d − 1) copula entropies by using Algorithm 6 6: For each copula function, calculate the 2 7: According to the log-likelihood function, select the adequate copula function from Table 2 for each edge 8: Build the corresponding graphical model 9: Following the structure of the graphical model, simulate the new population for each edge by using Algorithm 3

2 Other alternatives, such as goodness of fit tests, can be also considered. However, as stated in Brechmann

et al. (2012), the AIC is more reliable than using goodness of fit tests.

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

In this paper we assess the performance of two EDAs based on the chain and the tree graphical models. Both EDAs select the copula function for each chain link and tree branch. This modeling flexibility is not present in others EDAs based on copula functions, where just one copula function is previously chosen and used for modeling the dependence structure. It is worth noting that our proposed copula models are designed for modeling the search distribution in optimization tasks. Similarly, the R (R Core Team 2013) package copulaedas by González-Fernández and Soto (2013), provides tools for solving optimization problems by means of EDAs and copula functions. Although this package also contains implementations of graphical models, these probabilistic models are based on regular vines. For modeling purposes only, the interested reader is referred to the R package copula (Yan 2007).

5 Experiments In order to investigate the performance of the proposed EDAs, a benchmark of eleven functions for continuous optimization is utilized in the experiments. The conducted comparisons are made for contrasting the proposed algorithms with EDAs based on Gaussian distributions. Thus, we employ in the experiments four EDAs represented by the following notation – ChainGaussian Gaussian : A chain graphical model with Gaussian copula functions and Gaussian marginals. – ChainSelect Kernel : A chain graphical model with the copula selection procedure and Gaussian kernels as marginals. – TreeGaussian Gaussian : A tree graphical model with Gaussian copula functions and Gaussian marginals. – TreeSelect Kernel : A tree graphical model with the copula selection procedure and Gaussian kernels as marginals. Two marginal distributions are considered for the EDAs: a parametric model (Gaussian distribution) and a nonparametric model (Gaussian kernel). For EDAs with Gaussian kernels as marginal distributions, the bandwidth is estimated. Figure 4 shows the description of the test functions to be minimized. The benchmark test suite includes separable functions and non-separable functions, from which there are unimodal and multimodal functions. In addition, the search domains are symmetric and asymmetric. All test functions are scalable. We use test problems in 4, 8, 10, 12 and 16 dimensions. Each EDA is run 20 times for each problem. The population size is ten times the problem dimension. The maximum number of evaluations is 100,000. However, when convergence to a local minimum is detected the run is stopped. Any improvement less than 1 × 10−6 in 30 iterations is considered premature convergence. The goal is to reach the optimum with an error less than 1 × 10−4 .

123

R. Salinas-Gutiérrez et al.

Fig. 4 Names, mathematical definition, search domain, global minimum and properties of the test functions

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

5.1 Numerical results The results in dimensions 4, 8, and 16 for separable functions are reported in Tables 3 and 4, whereas the results for non-separable functions are reported in Tables 5 and 6. The tables report descriptive statistics for the fitness values reached in the all runs. The fitness value corresponds to the value of a test problem. For each algorithm and dimension, the minimum, median, mean, maximum, standard deviation and success rate are shown. The minimum (maximum) value reached is labelled best (worst). The success rate is the proportion of runs in which an algorithm found the global optimum. In order to properly compare the performance of the proposed EDAs, we conducted a hypothesis test to determine if the copula selection procedure can help in improving results. The hypotheses for the test are H0 : μa ≤ μb versus H1 : μa > Table 3 Descriptive results of the fitness for separable functions Algorithm

d

Best

Median

Mean

Worst

SD

Success rate

4

2.05E−5

7.70E−5

2.08E−1

2.89E+0

6.65E−1

0.70

8

2.21E−5

8.32E−5

3.00E−1

4.36E+0

9.91E−1

0.80

16

3.56E−5

6.95E−5

1.04E−2

2.01E−1

4.49E−2

0.90

4

1.47E−5

8.00E−5

3.49E+0

6.98E+1

1.56E+1

0.80 +

Cigar ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

8

3.85E−5

7.14E−5

7.00E−5*

9.98E−5

1.76E−5

1.00 +

16

3.70E−5

7.95E−5

7.58E−5

9.88E−5

1.72E−5

1.00 +

4

1.26E−5

7.61E−5

1.13E+0

8.64E+0

2.70E+0

0.75

8

3.19E−5

7.80E−5

1.88E−2

2.70E−1

6.24E−2

0.80

16

3.22E−5

9.03E−5

2.71E−1

3.33E+0

7.63E−1

0.70

4

7.20E−6

8.55E−5

1.60E+3

3.18E+4

7.10E+3

0.65

8

1.79E−5

7.42E−5

8.99E−4*

1.66E−2

3.70E−3

0.95 +

16

4.74E−5

7.85E−5

7.74E−5*

9.57E−5

1.43E−5

1.00 +

4

1.50E−5

7.04E−5

4.43E+2

8.84E+3

1.98E+3

0.70

8

3.26E−5

8.18E−5

4.86E−1

5.31E+0

1.39E+0

0.80

16

2.87E−5

7.15E−5

6.92E−5

9.99E−5

1.86E−5

1.00

4

1.81E−5

6.67E−5

3.49E+1

6.25E+2

1.40E+2

0.70

8

4.51E−5

7.72E−5

7.50E−5*

9.95E−5

1.92E−5

1.00 +

Cigar Tablet ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

16

4.47E−5

8.10E−5

7.86E−5

9.65E−5

1.59E−5

1.00

4

6.93E−6

9.40E−5

3.85E+0

5.08E+1

1.16E+1

0.55 0.95

8

3.28E−5

7.09E−5

7.45E+2

1.49E+4

3.33E+3

16

4.63E−5

8.38E−5

5.12E−1

1.02E+1

2.29E+0

0.90

4

1.55E−5

9.27E−5

1.65E+0

1.38E+1

3.86E+0

0.55

8

3.82E−5

7.74E−5

6.98E−5

9.01E−5

1.70E−5

1.00 +

16

5.94E−5

8.49E−5

8.14E−5

9.97E−5

1.36E−5

1.00 +

Ellipsoid

123

R. Salinas-Gutiérrez et al. Table 3 continued Algorithm ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

d

Best

Median

Mean

Worst

SD

Success rate

4

2.19E−5

8

2.49E−5

6.72E−5

9.78E+0

1.95E+2

4.36E+1

0.85

6.26E−5

3.24E+0

6.34E+1

1.42E+1

16

0.70

5.20E−5

8.49E−5

2.26E+1

2.57E+2

6.83E+1

4

0.75

2.28E−5

8.41E−5

5.43E−1

8.40E+0

1.90E+0

0.75

8

2.30E−5

7.00E−5

6.65E−5*

9.53E−5

1.96E−5

1.00 +

16

4.87E−5

8.34E−5

7.95E−5*

9.78E−5

1.59E−5

1.00 +

4

1.44E−5

7.82E−5

4.25E+1

8.41E+2

1.88E+2

0.65

8

4.54E−5

7.69E−5

1.08E−1

1.21E+0

3.29E−1

0.80

16

3.88E−5

9.33E−5

4.29E−1

3.33E+0

9.40E−1

0.55

4

1.26E−5

6.79E−5

6.19E−2*

6.76E−1

1.85E−1

0.80 +

8

3.86E−5

7.68E−5

7.51E−5*

9.76E−5

1.66E−5

1.00 +

16

4.32E−5

8.27E−5

7.93E−5**

9.95E−5

1.62E−5

1.00 +

* significantly better than the EDA with Gaussian copula, at α = 0.10 ** significantly better than the EDA with Gaussian copula, at α = 0.05 + that the EDA with copula selection procedure has greater success rate than the EDA with Gaussian copula

Table 4 Descriptive results of the fitness for separable functions Algorithm

d

Best

Median

Mean

Worst

SD

Success rate

4 8 16 4

3.83E−5 1.10E+1 5.74E+1 3.20E−6

2.39E+0 2.04E+1 7.40E+1 7.82E−5

2.31E+0 2.02E+1 7.41E+1 9.75E−1**

4.73E+0 2.77E+1 8.64E+1 8.55E+0

1.56E+0 4.40E+0 8.65E+0 2.28E+0

0.10 0.00 0.00 0.75 +

8

3.97E−5

1.55E+1

1.38E+1**

2.89E+1

1.07E+1

0.15 +

16

5.18E+1

7.42E+1

7.52E+1

9.39E+1

1.14E+1

0.00

4

1.25E−5

2.31E+0

2.55E+0

5.17E+0

1.49E+0

0.10

8

1.03E+1

1.96E+1

1.87E+1

2.59E+1

4.45E+0

0.00

16

6.16E+1

7.28E+1

7.29E+1

8.55E+1

6.57E+0

0.00

4

5.48E−7

8.62E−5

6.40E−1**

5.80E+0

1.35E+0

0.60 +

8

3.42E−5

1.32E+1

1.05E+1**

2.92E+1

1.04E+1

0.40 +

16

2.66E+1

8.22E+1

7.89E+1

1.04E+2

1.57E+1

0.00

4

5.90E−6

6.81E−5

1.37E−2

2.74E−1

6.11E−2

0.90

8

3.59E−5

7.39E−5

7.22E−5

9.99E−5

1.84E−5

1.00

16

4.25E−5

7.75E−5

7.62E−5

9.77E−5

1.53E−5

1.00

4

2.54E−5

5.10E−5

5.43E−5*

9.99E−5

2.16E−5

1.00 +

8

2.16E−5

7.84E−5

7.07E−5

9.65E−5

2.35E−5

1.00

16

5.29E−5

8.22E−5

8.00E−5

9.95E−5

1.29E−5

1.00

Rastrigin ChainGaussian Gaussian ChainSelect Kernel

TreeGaussian Gaussian

*TreeSelect Kernel

Sphere Model ChainGaussian Gaussian

ChainSelect Kernel

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms Table 4 continued Algorithm TreeGaussian Gaussian

TreeSelect Kernel

d

Best

Median

Mean

Worst

SD

Success rate

4

3.90E−6

8

3.19E−5

4.38E−5

5.08E−5

9.86E−5

3.08E−5

1.00

7.05E−5

6.74E−5

9.77E−5

1.75E−5

16

1.00

3.92E−5

7.09E−5

7.06E−5

9.57E−5

1.41E−5

4

1.00

8.63E−6

6.40E−5

9.46E−3

1.12E−1

2.78E−2

0.80

8

1.50E−5

5.64E−5

5.97E−5

9.75E−5

2.19E−5

1.00

16

4.89E−5

8.23E−5

8.10E−5

9.70E−5

1.24E−5

1.00

4

9.16E−6

6.35E−5

2.73E+0

4.20E+1

9.50E+0

0.75

8

2.49E−5

9.49E−5

7.83E−1

4.18E+0

1.30E+0

0.60

16

4.86E−5

9.60E−5

4.16E−1

4.91E+0

1.14E+0

0.55

4

2.89E−5

2.88E−3

8.60E−1

1.08E+1

2.59E+0

0.50

8

3.48E−5

6.88E−5

6.62E−5**

9.06E−5

1.52E−5

1.00 +

16

4.14E−5

8.05E−5

7.81E−5*

9.79E−5

1.64E−5

1.00 +

4

2.08E−5

7.61E−5

6.14E−1

1.03E+1

2.31E+0

0.70

8

2.15E−5

1.27E−2

1.51E+0

1.19E+1

2.89E+0

0.40

16

5.30E−5

1.87E−3

5.24E−1

3.80E+0

9.78E−1

0.50

4

1.79E−5

8.71E−5

1.02E−1

1.69E+0

3.76E−1

0.65

8

3.18E−5

7.15E−5

6.93E−5**

9.73E−5

1.42E−5

1.00 +

16

5.83E−5

7.99E−5

7.87E−5**

9.78E−5

1.17E−5

1.00 +

Two Axes ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

* significantly better than the EDA with Gaussian copula, at α = 0.10 ** significantly better than the EDA with Gaussian copula, at α = 0.05 + that the EDA with copula selection procedure has greater success rate than the EDA with Gaussian copula

μb , where μa stands for the fitness average of an EDA based on Gaussian copTable 5 Descriptive results of the fitness for non-separable functions Algorithm

d

Best

Median

Mean

Worst

SD

Success rate

4

2.15E−5

7.34E−5

9.11E−3

1.65E−1

3.68E−2

0.90

8

5.56E−5

8.48E−5

8.11E−5

9.94E−5

1.51E−5

1.00

16

7.27E−5

8.85E−5

8.82E−5

9.94E−5

7.43E−6

1.00

4

2.80E−5

7.04E−5

6.99E−5*

9.74E−5

2.12E−5

1.00 + 1.00

Ackley ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

8

6.40E−5

8.92E−5

8.49E−5

9.87E−5

1.15E−5

16

7.59E−5

9.20E−5

9.00E−5

9.88E−5

7.27E−6

1.00

4

3.72E−5

6.25E−5

6.71E−5

1.49E−4

2.63E−5

0.90

8

5.03E−5

7.88E−5

7.91E−5

9.88E−5

1.52E−5

1.00

16

7.29E−5

9.24E−5

8.85E−5

9.92E−5

8.80E−6

1.00

4

2.57E−5

7.16E−5

1.19E−2

1.94E−1

4.36E−2

0.85

8

4.81E−5

8.72E−5

8.39E−5

9.82E−5

1.24E−5

1.00

16

6.13E−5

9.13E−5

8.82E−5

9.99E−5

1.04E−5

1.00

123

R. Salinas-Gutiérrez et al. Table 5 continued Algorithm

d

Best

Median

Mean

Worst

SD

Success rate

4

3.43E−2

8

1.81E−1

1.01E−1

1.10E−1

2.20E−1

5.28E−2

0.00

3.67E−1

3.69E−1

5.53E−1

9.73E−2

16

0.00

3.58E−5

8.18E−5

7.81E−5

9.92E−5

1.79E−5

1.00

4

2.96E−2

1.02E−1

1.19E−1

2.83E−1

6.68E−2

0.00

8

7.40E−3

2.03E−1

2.21E−1**

5.22E−1

1.31E−1

0.00

16

6.51E−5

8.02E−5

8.15E−5

9.92E−5

9.65E−6

1.00

4

4.79E−2

1.26E−1

1.26E−1

2.21E−1

5.21E−2

0.00

8

1.58E−1

3.80E−1

3.78E−1

5.49E−1

1.10E−1

0.00

16

4.41E−5

7.60E−5

7.68E−5

9.70E−5

1.30E−5

1.00

4

2.90E−2

8.92E−2

1.02E−1*

2.06E−1

5.05E−2

0.00

8

5.40E−5

2.45E−1

2.53E−1**

5.34E−1

1.52E−1

0.05 +

16

2.93E−5

7.59E−5

7.50E−5

9.97E−5

1.80E−5

1.00

4

2.53E−5

6.00E−3

1.49E−1

9.44E−1

2.76E−1

0.40

8

9.94E−5

4.89E−2

2.82E−1

2.04E+0

5.92E−1

0.10

16

1.38E−2

3.94E−1

1.07E+0

5.60E+0

1.56E+0

0.00

4

1.84E−5

9.74E−5

1.43E−2**

1.41E−1

3.58E−2

0.60 +

8

4.95E−5

1.01E−4

8.38E−3**

7.25E−2

1.83E−2

0.50 +

16

9.54E−5

2.77E−4

2.14E−2**

2.21E−1

5.08E−2

0.40 +

4

5.82E−6

7.78E−5

1.02E−2

5.44E−2

1.78E−2

0.55

8

9.65E−5

1.94E−2

4.50E−1

3.44E+0

8.87E−1

0.05

16

3.66E−2

4.34E−1

1.15E+0

6.78E+0

1.81E+0

0.00

4

3.03E−5

8.15E−5

7.26E−2

1.39E+0

3.10E−1

0.65 +

8

5.21E−5

9.52E−5

3.84E−4**

2.66E−3

7.50E−4

0.75 +

16

9.47E−5

2.89E−4

4.85E−3**

6.74E−2

1.54E−2

0.40 +

Griewangk ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

Schwefel 1.2 ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

* significantly better than the EDA with Gaussian copula, at α = 0.10 ** significantly better than the EDA with Gaussian copula, at α = 0.05 + that the EDA with copula selection procedure has greater success rate than the EDA with Gaussian copula

ula and μb stands for the fitness average of an EDA based on copula selection. The statistical comparisons are for the algorithms with the same graphical model. The hypothesis test is based on a bootstrap method. When a null hypothesis can be rejected, the corresponding average μb is marked with an asterix (*). Besides the average fitness, the success rate is another performance measure that can help in the comparisons of the algorithms. Figures 5 and 6 show respectively the success rate in each dimension for separable and non-separable functions. If the success rate of an EDA based on copula selection is greater than the success rate of the corresponding algorithm without copula selection, it is marked with plus sign (+).

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms Table 6 Descriptive results of the fitness for non-separable functions Algorithm Trid ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

d

Best

Median

Mean

Worst

SD

Success rate

4

−1.60E+1

−1.60E+1

−1.58E+1

−1.17E+1

9.63E−1

0.65

8

−1.12E+2

−1.12E+2

−1.11E+2

−1.04E+2

2.14E+0

0.65

16

−8.00E+2

−8.00E+2

−7.92E+2

−6.92E+2

2.41E+1

0.55

4

−1.60E+1

−1.60E+1

−1.60E+1

−1.56E+1

9.51E−2

0.80 +

8

−1.12E+2

−1.12E+2

−1.12E+2*

−1.12E+2

7.85E−4

0.90 +

16

−8.00E+2

−8.00E+2

−8.00E+2*

−7.99E+2

1.26E−1

0.60 +

4

−1.60E+1

−1.60E+1

−1.59E+1

−1.48E+1

2.68E−1

0.65

8

−1.12E+2

−1.12E+2

−1.12E+2

−1.09E+2

8.04E−1

0.55

16

−8.00E+2

−8.00E+2

−7.99E+2

−7.93E+2

1.68E+0

0.45

4

−1.60E+1

−1.60E+1

−1.60E+1*

−1.60E+1

1.86E−4

0.90 +

8

−1.12E+2

−1.12E+2

−1.12E+2

−1.10E+2

6.22E−1

0.65 +

16

−8.00E+2

−8.00E+2

−8.00E+2

−7.97E+2

9.42E−1

0.60 +

4

3.84E−5

9.59E−5

1.26E−1

2.13E+0

4.78E−1

0.55

8

8.76E−5

1.36E−2

1.23E−1

9.53E−1

2.38E−1

0.05

16

8.52E−5

2.56E−3

7.51E−3

3.13E−2

9.44E−3

0.20

4

1.05E−5

6.76E−5

4.47E−3*

8.44E−2

1.88E−2

0.90 +

Zakharov ChainGaussian Gaussian

ChainSelect Kernel

TreeGaussian Gaussian

TreeSelect Kernel

8

4.18E−5

9.80E−5

2.51E−3**

4.62E−2

1.03E−2

0.80 +

16

5.93E−5

9.79E−5

1.65E+1

1.45E+2

3.66E+1

0.70 +

4

1.99E−5

7.41E−4

1.87E−2

1.35E−1

3.90E−2

0.50

8

9.92E−5

8.63E−3

6.46E−2

4.05E−1

1.13E−1

0.05

16

8.52E−5

3.34E−2

3.56E−1

4.80E+0

1.06E+0

0.15

4

1.58E−5

7.83E−5

2.10E−1

2.60E+0

6.55E−1

0.65 +

8

3.86E−5

9.87E−5

1.59E−2*

3.11E−1

6.94E−2

0.85 +

16

8.07E−5

9.05E−5

1.40E+1

1.22E+2

3.65E+1

0.80 +

* significantly better than the EDA with Gaussian copula, at α = 0.10 ** significantly better than the EDA with Gaussian copula, at α = 0.05 + that the EDA with copula selection procedure has greater success rate than the EDA with Gaussian copula

For every function six comparisons are shown in the Tables 3, 4, 5, and 6, that is, two algorithms times three dimensions. The tables show a total of 66 comparisons. Out of the 36 comparisons for the separable functions, the EDA with copula selection excels in 22 cases, it has similar performance in 9 cases, and it is outperformed in only 5 cases. Similarly, out of the 30 comparisons for the non separable functions, the EDA with copula selection excels in 20 cases, it has similar performance in 9 cases, and it is outperformed in only one case. This could imply that the EDAs based on a copula selection procedure are more adequate than algorithms with a fixed dependence function for non-separable optimization problems.

123

R. Salinas-Gutiérrez et al.

Separable Functions Chains

Chains

1

1

0.5

0.5

0

0 Trees

Trees

1

1

0.5

0.5

0

0 4

8

10

12

16

4

Cigar

8

10

12

16

Cigar Tablet

Chains

Chains

1

1

0.5

0.5

0

0 Trees

Trees

1

1

0.5

0.5

0

0 4

8

10

12

16

4

8

Ellipsoid

10

12

16

Rastrigin

Chains

Chains

1

1

0.5

0.5

0

0 Trees

Trees

1

1

0.5

0.5

0

0 4

8

10

12

Sphere model

16

4

8

10

12

16

Two Axes

Fig. 5 Success rate in each dimension for separable functions. The horizontal axis represents the dimension problem and the vertical axis represents the success rate. The solid line is used for the EDAs based on copula selection and the dashed line for the EDAs based on the Gaussian copula

5.2 Discussion 5.2.1 Separable functions In general, the unimodal and separable functions are not difficult to solve. The algorithms have a good performance in solving all the functions with exception of the multimodal Rastrigin function. However, the statistical tests and the success rates show that the EDAs based on copula selection can achieve and outperform the results founded by the algorithms based on Gaussian copula. The success rates for EDAs based on copula selection do not decrease with the dimension problem in all separable functions, except the Rastrigin function. Moreover,

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms

Non-separable Functions Chains

Chains

1

1

0.5

0.5

0

0 Trees

Trees

1

1

0.5

0.5

0

0 4

8

10

12

16

4

Ackley

8

10

12

16

Griewangk

Chains

Chains

1

1

0.5

0.5

0

0 Trees

Trees

1

1

0.5

0.5

0

0 4

8

10

12

16

Schwefel 1.2

4

8

10

12

16

Trid

Chains

1 0.5 0 Trees

1 0.5 0 4

8

10

12

16

Zakharov Fig. 6 Success rate in each dimension for non-separable functions. The horizontal axis represents the dimension problem and the vertical axis represents the success rate. The solid line is used for the EDAs based on copula selection and the dashed line for the EDAs based on the Gaussian copula

their success rate in functions Cigar, Cigar Tablet, Ellipsoid, Sphere model, and Two Axes is always 100 % in dimension 16.

5.2.2 Non-separable functions It is known that this kind of test functions have associations or dependencies among their variables, thus, they are difficult to solve. The performance of all algorithms in the Ackley and Griewangk functions shows that their success rate does not decrease its value with the dimension problem. For the Schwefel 1.2, Trid, and Zakharov functions, all the algorithms decrease their success rate with the dimensionality.

123

R. Salinas-Gutiérrez et al. Table 7 Descriptive results of the fitness for separable functions by using EMNA d

Best

Median

Mean

Worst

SD

Success rate

4

3.33E−5

3.17E+0

9.80E+0

5.19E+1

1.36E+1

0.20

8

1.56E−4

2.35E+2

3.96E+4

5.14E+5

1.19E+5

0.00

16

5.99E+3

6.24E+5

8.87E+5

3.81E+6

1.01E+6

0.00

Cigar

Cigar Tablet 4

1.33E−5

2.84E−2

1.30E+0

1.09E+1

2.68E+0

0.25

8

5.48E−5

2.24E+0

5.08E+0

2.62E+1

7.75E+0

0.10

16

7.80E−1

1.05E+1

9.68E+0

1.92E+1

4.45E+0

0.00

4

1.52E−5

2.09E+0

4.20E+1

4.67E+2

1.06E+2

0.20

8

1.73E−4

1.08E+2

4.57E+2

5.52E+3

1.24E+3

0.00

16

1.39E+2

8.79E+2

1.30E+3

4.62E+3

1.11E+3

0.00

4

9.96E−5

2.65E+0

3.21E+0

6.85E+0

1.64E+0

0.05

8

1.06E+1

2.08E+1

2.05E+1

2.73E+1

4.50E+0

0.00

16

5.12E+1

7.21E+1

7.11E+1

8.27E+1

7.62E+0

0.00

Ellipsoid

Rastrigin

Sphere Model 4

8.80E−6

7.69E−5

1.81E+1

3.61E+2

8.08E+1

0.70

8

7.72E−5

1.47E+0

3.71E+1

4.64E+2

1.05E+2

0.10

16

1.17E+0

3.15E+1

1.17E+2

5.88E+2

1.81E+2

0.00

4

1.71E−5

1.42E+0

3.64E+0

2.09E+1

5.52E+0

0.15

8

6.31E−2

7.59E+0

1.09E+1

4.73E+1

1.15E+1

0.00

16

3.75E+0

4.33E+1

7.15E+1

5.91E+2

1.24E+2

0.00

Two Axes

5.2.3 Comparison with EMNA A well established EDA for solving optimization problems with continuous variables is the EMNA Larrañaga and Lozano (2002). This algorithm is based on the multivariate Normal distribution as probabilistic model. The EMNA is also used for solving the test problems of Fig. 4. The results for each optimization problem are shown in Tables 7 and 8. These additional results allow us to compare the performance of EDAs based on copula functions, Tables 3, 4, 5, and 6 with a standard EDA such as the EMNA. For most of the test problems in all dimensions, with exception of the functions Schwefel 1.2 and Zakharov, the EMNA was outperformed by the EDAs based on copula functions. 6 Conclusions In this paper a copula selection procedure for continuous EDAs has been introduced. The implementation of this procedure is shown in two well known graphical models.

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms Table 8 Descriptive results of the fitness for non-separable functions by using EMNA d

Best

Median

Mean

Worst

SD

Success rate

Ackley 4

3.44E−5

9.57E−5

1.01E−1

1.84E+0

4.10E−1

0.60

8

8.62E−5

2.09E−2

9.96E−2

5.74E−1

1.52E−1

0.05

16

3.34E−2

1.10E−1

1.98E−1

7.72E−1

2.02E−1

0.00

Griewangk 4

3.76E−2

1.13E−1

1.13E−1

1.99E−1

4.12E−2

0.00

8

9.90E−3

3.11E−1

2.97E−1

5.73E−1

1.53E−1

0.00

16

1.09E−2

4.66E−1

4.68E−1

1.04E+0

3.48E−1

0.00

Schwefel 1.2 4

1.79E−5

7.10E−5

2.58E−2

2.25E−1

6.23E−2

0.70

8

7.05E−5

5.10E−3

8.57E−1

6.67E+0

1.86E+0

0.20

16

8.57E−5

1.10E+0

2.27E+0

2.04E+1

4.43E+0

0.05

4

−1.60E+1

−1.60E+1

−1.60E+1

−1.59E+1

2.62E−2

0.55

8

−1.12E+2

−1.12E+2

−1.08E+2

−5.87E+1

1.20E+1

0.10

16

−7.92E+2

−6.80E+2

−6.46E+2

−4.32E+2

1.17E+2

0.00

Trid

Zakharov 4

2.13E−5

6.65E−5

3.66E−1

6.84E+0

1.53E+0

0.65

8

5.17E−5

7.73E−3

6.66E−1

4.78E+0

1.34E+0

0.20

16

3.98E−3

1.11E+0

3.20E+0

2.59E+1

6.43E+0

0.00

According to the numerical experiments the selection of a copula function for modelling the dependence structure can help achieving better fitness values. This means that dependencies between decision variables must be modeled adequately in order to get good solutions. The EDA based on a chain graphical model and the EDA based on a tree graphical model, both with a copula selection procedure, have a better performance than EDAs based on multivariate Gaussian distributions. The success rate already indicates a better performance of the algorithms adapted with copula selection in higher dimension. Although the statistical comparisons show similar performances of both algorithms in reaching the function optima, the success rate clearly indicates that EDAs with copula selection are more robust since they reach the optima in a larger number of experiments. Nonetheless, more experiments are necessary with different graphical models in order to identify where the copula functions have a clear advantage to EDAs. Given that all the algorithms presented in this paper estimate their probabilistic parameters just by means of maximum likelihood, the immediate research work will focus on the design of algorithms with diversity maintenance in the population, or similar strategies, for enhancing the performance of copula functions in EDAs. Acknowledgments The first author acknowledges support from the Center for Research in Mathematics and from the Programa de Mejoramiento del Profesorado (PROMEP) of México.

123

R. Salinas-Gutiérrez et al.

References Aas K, Czado C, Frigessi A, Bekken H (2009) Pair-copula constructions of multiple dependence. Insur Math Econ 44(2):182–198 Abreu G, Labouriau R, Edwards D (2010) High-dimensional graphical model search with the gRapHD R package. J Stat Softw 37(1):1–18 Baluja S (1994) Population-based incremental learning: a Method for integrating genetic search based function optimization and competitive learning. Technical Report CMU-CS-94-163, Carnegie Mellon University, Pittsburgh, PA Baluja S, Davies S (1997) Using optimal dependency-trees for combinatorial optimization: learning the structure of the search space. In: Fisher D (ed) Proceedings of the fourteenth international conference on machine learning. Morgan Kaufmann, San Francisco, CA, pp 30–38 Bedford T, Cooke RM (2002) Vines—a new graphical model for dependent random variables. Ann Stat 30(4):1031–1068 Bosman P, Grahl J, Thierens D (2008) Enhancing the performance of maximum-likelihood Gaussian EDAs using anticipated mean shift. In: Rudolph G, Jansen T, Lucas S, Poloni C, Beume N (eds) Parallel problem solving from nature—PPSN X, lecture notes in computer science, vol 5199. Springer, Berlin, pp 133–143 Brechmann EC, Czado C, Aas K (2012) Truncated regular vines in high dimensions with application to financial data. Can J Stat 40(1):68–85 Cherubini U, Luciano E, Vecchiato W (2004) Copula methods in Finance. Wiley, Chichester Chow C, Liu C (1968) Approximating discrete probability distributions with dependence trees. IEEE Trans Inf Theory 14(3):462–467 Cormen T, Leiserson C, Rivest R, Stein C (2009) Introduction to algorithms, 3rd edn. The MIT Press, Cambridge, MA Cover T, Thomas J (1991) Elements of information theory, 1st edn. Wiley, New York, NY Cuesta-Infante A, Santana R, Hidalgo J, Bielza C, Larrañaga P (2010) Bivariate empirical and n-variate Archimedean copulas in estimation of distribution algorithms. In: WCCI 2010 IEEE world congress on computational intelligence. Barcelona, pp 1355–1362 Davy M, Doucet A (2003) Copulas: a new insight into positive time-frequency distributions. Signal Process Lett IEEE 10(7):215–218. doi:10.1109/LSP.2003.811636 De Bonet J, Isbell C, Viola P (1997) MIMIC: finding optima by estimating probability densities. Advances in neural information processing systems, vol 9. The MIT Press, Cambridge, MA, pp 424–430 Gao Y (2009) Multivariate estimation of distribution algorithm with laplace transform Archimedean copula. In: Hu W, Li X (eds) 2009 International conference on information engineering and computer science, ICIECS 2009. Wuhan, China González-Fernández Y, Soto M (2013) copulaedas: estimation of distribution algorithms based on copulas. http://CRAN.R-project.org/package=copulaedas, R package version 1.3.0 Harik G, Lobo F, Goldberg D (1998) The compact genetic algorithm. In: Proceedings of the IEEE conference on evolutionary computation. Anchorage, AK, pp 523–528 Igel C, Suttorp T, Hansen N (2006) A computational efficient covariance matrix update and a (1+1)CMA for evolution strategies. In: Proceedings of the 8th annual conference on genetic and evolutionary computation. ACM, New York, NY, GECCO ’06, pp 453–460 Joe H (1997) Multivariate models and dependence concepts. Chapman & Hall, London Kim G, Silvapulle MJ, Silvapulle P (2007) Comparison of semiparametric and parametric methods for estimating copulas. Comput Stat Data Anal 51(6):2836–2850 Koller D, Friedman N (2009) Probabilistic graphical models: principles and techniques. The MIT Press, Cambridge, MA Kurowicka D, Cooke R (2006) Uncertainty analysis. Wiley, Chichester Larrañaga P, Lozano J (eds) (2002) Estimation of distribution algorithms: a new tool for evolutionary computation. Genetic algorithms and evolutionary computation. Kluwer, Norwell, MA Mühlenbein H (1998) The equation for response to selection and its use for prediction. Evolut Comput 5(3):303–346 Mühlenbein H, Paaß G (1996) From recombination of genes to the estimation of distributions I. Binary parameters. In: Voigt H, Ebeling W, Rechenberg I, Schwefel H (eds) Parallel problem solving from nature—PPSN IV, Lecture Notes in Computer Science, vol 1141. Springer, Berlin, pp 178–187 Nelsen R (2006) An introduction to copulas, 2nd edn. Springer, New York, NY

123

Copula selection for graphical models in continuous Estimation of Distribution Algorithms Pelikan M, Mühlenbein H (1999) The bivariate marginal distribution algorithm. In: Roy R, Furuhashi T, Chawdhry P (eds) Advances in soft computing—engineering design and manufacturing. Springer, London, pp 521–535 Pelikan M, Goldberg D, Cantú-Paz E (1999) BOA: the Bayesian optimization algorithm. In: Banzhaf W, Daida J, Eiben A, Garzon M, Honavar V, Jakiela M, Smith R (eds) Proceedings of the genetic and evolutionary computation conference GECCO-99, vol 1. Morgan Kaufmann PublishersOrlando, FL, pp 525–532 R Core Team (2013) R: a language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria, http://www.R-project.org Salinas-Gutiérrez R, Hernández-Aguirre A, Villa-Diharce E (2009) Using copulas in estimation of distribution algorithms. In: Hernández Aguirre A, Monroy Borja R, Reyes García C (eds) MICAI 2009: advances in artificial intelligence, lecture notes in computer science, vol 5845. Springer, Berlin, pp 658–668 Salinas-Gutiérrez R, Hernández-Aguirre A, Villa-Diharce E (2010) D-vine EDA: a new estimation of distribution algorithm based on regular vines. In: GECCO ’10: proceedings of the 12th annual conference on genetic and evolutionary computation. ACM, New York, NY, pp 359–366 Sklar A (1959) Fonctions de répartition à n dimensions et leurs marges. Publications de l’Institut de Statistique de l’Université de Paris 8:229–231 Soto M, Ochoa A, Acid S, de Campos L (1999) Introducing the polytree approximation of distribution algorithm. In: Ochoa A, Soto M, Santana R (eds) Second international symposium on artificial intelligence. Adaptive systems. CIMAF99, Academia, La Habana, pp 360–367 Trivedi P, Zimmer D (2007) Copula modeling: an introduction for practitioners, foundations and trends® in econometrics, vol 1. Now Publishers, New York, NY Wang L, Zeng J (2010) Estimation of distribution algorithm based on copula theory. In: Chen Y (ed) Exploitation of linkage learning in evolutionary algorithms, adaptation, learning, and optimization, vol 3. Springer, Berlin, pp 139–162 Wang L, Wang Y, Zeng J, Hong Y (2010a) An estimation of distribution algorithm based on clayton copula and empirical margins. In: Li K, Li X, Ma S, Irwin G (eds) Life system modeling and intelligent computing. Springer, Berlin, pp 82–88 Wang L, Zeng J, Hong Y, Guo X (2010b) Copula estimation of distribution algorithm sampling from clayton copula. J Comput Inf Syst 6(7):2431–2440 Weiß G (2011) Copula parameter estimation by maximum-likelihood and minimum-distance estimators: a simulation study. Comput Stat 26(1):31–54 Yan J (2007) Enjoy the joy of copulas: with a package copula. J Stat Softw 21(4):1–21

123