HD-DDS

2 downloads 0 Views 437KB Size Report
Hybrid Discrete Dynamically Dimensioned Search (HD-DDS) Algorithm for. 1 ...... Intelligence, Vancouver, Canada, July 16-21, Proceedings on DVD. 17. Cunha ...
Tolson, B. A., M. Asadzadeh, H. R. Maier, and A. Zecchin (2009), Hybrid discrete dynamically dimensioned search (HD-DDS) algorithm for water distribution system design optimization, Water Resources Research, 45, W12416, doi:10.1029/2008WR007673. 1

Hybrid Discrete Dynamically Dimensioned Search (HD-DDS) Algorithm for

2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19

Water Distribution System Design Optimization

20

Abstract

Bryan A. Tolson Department of Civil Engineering, University of Waterloo, 200 University Avenue West, Waterloo, Ontario, N2L 3G1, Canada ([email protected])

Masoud Asadzadeh Department of Civil Engineering, University of Waterloo, 200 University Avenue West, Waterloo, Ontario, N2L 3G1, Canada ([email protected])

Aaron Zecchin School of Civil, Environmental and Mining Engineering, University of Adelaide, Adelaide, SA 5005 Australia ([email protected])

Holger R. Maier School of Civil, Environmental and Mining Engineering, University of Adelaide, Adelaide, SA 5005 Australia ([email protected])

21

The Dynamically Dimensioned Search (DDS) continuous global optimization algorithm by

22

Tolson and Shoemaker [2007] is modified to solve discrete, single-objective, constrained Water

23

Distribution System (WDS) design problems. The new global optimization algorithm for WDS

24

optimization is called Hybrid Discrete Dynamically Dimensioned Search (HD-DDS) and

25

combines two local search heuristics with a discrete DDS search strategy adapted from the

26

continuous DDS algorithm. The main advantage of the HD-DDS algorithm compared with other

27

heuristic global optimization algorithms, such as genetic and ant colony algorithms, is that its

28

searching capability (i.e. the ability to find near globally optimal solutions) is as good, if not

29

better, while being significantly more computationally efficient. The algorithm’s computational

30

efficiency is due to a number of factors, including the fact that it is not a population-based

31

algorithm and only requires computationally expensive hydraulic simulations to be conducted for

32

a fraction of the solutions evaluated. This paper introduces and evaluates the algorithm by

33

comparing its performance with that of three other algorithms (specific versions of the Genetic

34

Algorithm, Ant Colony Optimization, and Particle Swarm Optimization) on four WDS case

35

studies (21- to 454-dimensional optimization problems) on which these algorithms have been 1

1

found to perform well. The results obtained indicate that the HD-DDS algorithm outperforms the

2

state-of-the-art existing algorithms in terms of searching ability and computational efficiency. In

3

addition, the algorithm is easier to use, as it does not require any parameter tuning and

4

automatically adjusts its search to find good solutions given the available computational budget.

5 6

Keywords: Water Distribution Systems; Discrete Optimization; Global optimization; Heuristic

7

optimization; Constraint-handling

8 9

1.

Introduction

10 11

The optimal design and rehabilitation of Water Distribution Systems (WDSs) is an important

12

research area as improved optimization methods save substantial infrastructure capital and

13

operational costs. Historically, traditional optimization methods such as linear programming

14

[Schaake and Lai, 1969; Alperovits and Shamir, 1977; Bhave and Sonak, 1992], nonlinear two

15

phase decomposition methods [Fujiwara and Khang, 1990; Eiger et al., 1994] and nonlinear

16

programming [Varma et al., 1997] have been applied to a continuous version of the WDS

17

optimization problem. These methods are sophisticated in terms of their use of the fundamental

18

hydraulic equations to recast the form of the optimization problem and to yield gradient and

19

hessian information, however, their inability to restrict the search space to discrete pipe sizes is a

20

significant practical limitation [Cunha and Sousa, 1999]. Reca and Martinez [2006] provide a

21

more detailed review of classical optimization techniques as applied to WDS optimization

22

problems.

23

The majority of current single-objective WDS optimization literature report using heuristic

24

global optimization algorithms, including evolutionary algorithms, with great success. Genetic

25

Algorithms (GAs) are probably the most well known combinatorial evolutionary algorithm.

26

Example applications of GAs to WDS optimization include Simpson et al. [1994], Savic and

27

Walters [1997], Wu et al. [2001] and Tolson et al. [2004]. Ant Colony Optimization (ACO) 2

1

algorithms have also received attention in the recent WDS literature [Maier et al., 2003; Zecchin

2

et al. 2006; Zecchin et al. 2007]. Other new and promising approaches applied to WDS

3

optimization include the Shuffled Frog Leaping Algorithm (SFLA) in Eusuff and Lansey [2003],

4

the Harmony Search (HS) Algorithm in Geem [2006], the Cross Entropy (CE) method in

5

Perelman and Ostfeld [2007], Particle Swarm Optimization (PSO) in Suribabu and Neelakantan

6

[2006] and another PSO variant in Montalvo et al. [2008]. Additionally, more traditional heuristic

7

search strategies that are not population-based, such as Simulated Annealing [Cunha and Sousa,

8

1999] and Tabu Search [Lippai et al., 1999; Cunha and Ribeiro, 2004] continue to be applied to

9

some WDS optimization problems.

10

The specific WDS optimization problem we address is to determine the pipe diameters from

11

a discrete set of available options such that the total pipe material cost is minimized and pressure

12

constraints are met across the network. All other network characteristics are known. This is a

13

classical WDS design problem that the majority of the above WDS optimization references also

14

solve.

15

One of the major problems associated with the use of heuristic global optimization

16

algorithms is that their performance, both in terms of computational efficiency and their ability to

17

find near globally optimal solutions, can be affected significantly by the settings of a number of

18

parameters that control their searching behavior (e.g. population size, probability of mutation,

19

probability of crossover in the case of GAs), as well as penalty functions that are commonly used

20

to account for system constraints. In accordance with the No Free Lunch Theorem [Wolpert and

21

MacReady, 1997], the set of parameters that results in optimal performance will vary with the

22

characteristics of each optimization problem. Consequently, values of these parameters are

23

generally obtained by trial and error for different case studies.

24

The algorithm parameter characteristics for eight recent WDS optimization studies are

25

summarized in Table 1. As can be seen, the reported number of total parameters (algorithm +

26

penalty) in these eight algorithms ranges from three to eight, and seven of these eight studies

27

report that a subset of algorithm parameters were either experimented with or modified for each 3

1

of the case studies they were applied to. However, such algorithm parameter setting experiments

2

can be extremely computationally expensive, as many of the parameters are dependent. In

3

addition, this problem is exacerbated for problems with long WDS simulation times, particularly

4

those requiring extended period simulations (e.g. when water quality considerations are

5

important). In addition to requiring a large amount of computational time, there is no well

6

accepted methodology for conducting these parameter setting experiments, which are therefore

7

typically implemented on an ad-hoc basis. For example, the methodology used to determine the

8

the optimal parameter settings for many of the studies in Table 1 is not described.

9

The issue of heuristic optimization parameter tuning has been investigated in a number of

10

recent studies. In relation to GAs, one approach is to self adapt the parameters as part of the

11

optimization procedure itself [e.g. Srinivasa et al., 2007]. Alternative approaches are based on

12

parameterless GA calibration methodologies [e.g. Lobo and Goldberg, 2004, Minsker, 2005] and

13

GA convergence behavior associated with genetic drift [e.g. Rogers and Prugel-Bennett, 1999,

14

Gibbs et al., 2008]. Gibbs et al. [2009] compared the performance of the above approaches on

15

the Cherry Hills-Brushy Plains WDS optimization problem [Bocelli et al., 1998] and found that

16

the approach based on genetic drift performed best overall. In relation to ACO, Zecchin et al.

17

[2005] used a mixture of theoretical and sensitivity analyses to derive expressions for seven ACO

18

parameters, which have been shown to perform well for a number of benchmark WDS

19

optimization problems [Zecchin et al., 2007]. One approach for eliminating penalty parameters

20

or multipliers in single objective WDS optimization problems is to convert hydraulic constraints

21

into objectives and therefore solve a multi-objective optimization problem without hydraulic

22

constraints [e.g. Wu and Simpson, 2002; Farmani et al., 2005].

23

Despite these efforts, common practice in many current heuristic optimization methods for

24

WDS design (see Table 1) still involves case study specific experimentation for tuning algorithm

25

parameters. As discussed previously, such experimentation is undesirable, as it has the potential

26

to increase computational effort significantly.

27

experimentation is sufficient and what impact limited experimentation has on algorithm

In addition, it is unclear how much

4

1

performance for a particular problem. In order to address these issues, the Hybrid Discrete

2

Dynamically Dimensioned Search (HD-DDS) algorithm is introduced in this paper. Not only is

3

the performance of HD-DDS reliant on only one parameter, but the value of this parameter does

4

not have to be adjusted for different case studies. This is in contrast to existing heuristic

5

optimization methods (Table 1).

6

Dimensioned Search (DDS) algorithm introduced by Tolson and Shoemaker [2007] for

7

continuous optimization problems and can be used for solving constrained, single-objective

8

combinatorial WDS optimization.

The HD-DDS algorithm builds on the Dynamically

9

The remainder of the paper is organized as follows. Section 2.1 describes the HD-DDS

10

algorithm in detail. The optimization algorithms with which HD-DDS is compared in order to

11

evaluate its utility are outlined in section 2.2, and the WDS benchmarks to which it is applied

12

(ranging from 21- to 454-dimensional problems) are introduced in section 2.3. Results are

13

presented in section 3 and are followed by a discussion in section 4 and conclusions in section 5.

14 15

2.

Methodology

16

2.1. Components of the Hybrid Discrete Dynamically Dimensioned Search Algorithm

17

The Hybrid Discrete Dynamically Dimensioned Search (HD-DDS) Algorithm for WDS

18

design optimization is described in the following sections. The HD-DDS algorithm utilizes global

19

and local search strategies, as such a hybrid approach has been shown to be successful previously

20

[e.g. Broad et al., 2006]. Sections 2.1.1 through 2.1.4 describe how each strategy functions

21

independently and then section 2.1.5 describes how they are combined to form the overall HD-

22

DDS algorithm.

23

2.1.1.

Discrete Dynamically Dimensioned Search Algorithm

24

The discrete Dynamically Dimensioned Search (discrete DDS) algorithm is the most

25

important component of the HD-DDS algorithm and is a discrete adaptation of the DDS 5

1

algorithm recently introduced by Tolson and Shoemaker [2007] for continuous optimization

2

problems. DDS was designed as a simple and parsimonious algorithm (it has only one algorithm

3

parameter) to solve computationally intensive environmental simulation model automatic

4

calibration problems. One DDS design goal was to have the algorithm automatically adjust and

5

exhibit good performance within the user’s timeframe for optimization (maximum number of

6

objective function evaluations) rather than require the user to modify and/or experiment with

7

algorithm parameters to match their timeframe. A related DDS design goal was to eliminate the

8

need for algorithm parameter adjustment when the case study or number of decision variables

9

changes. While it is acknowledged that this is unlikely to result in the identification of globally

10

optimal solutions, the DDS algorithm is simple to use in practice, while being able to consistently

11

find near globally optimal solutions. In fact, Tolson and Shoemaker [2007] demonstrate better

12

overall performance of DDS relative to other benchmark automatic calibration algorithms on

13

optimization problems ranging from 6- to 30-dimensions with 1000 to 10,000 objective function

14

evaluations per optimization trial while using the same DDS algorithm parameter value.

15

The parsimonious nature of DDS provides an attractive alternative for discrete WDS

16

optimization problems given the very recent set of single objective WDS optimization algorithms

17

reviewed in section 1 (see Table 1), all of which have from three to eight algorithm parameters of

18

which a subset is usually modified and even optimized for different case studies. The discrete

19

DDS algorithm is identical to the original DDS algorithm except for two modifications. The first

20

of these enables the proposed algorithm to sample discrete valued candidate solutions, whereas

21

the second is the addition of a new algorithm stopping criterion. The paragraphs below describe

22

the DDS algorithm (largely from Tolson and Shoemaker [2007]) and are followed by a

23

description of the two modifications that distinguish discrete DDS from DDS.

24

The DDS algorithm is a simple, stochastic, single-solution based, heuristic, global search

25

algorithm that was developed for the purpose of finding a good approximation of the globally

26

optimal solution within a specified maximum number of objective function evaluations. The

27

algorithm is designed to scale the search to a user-specified number of maximum objective 6

1

function evaluations. In short, DDS searches globally at the start of the search, transitioning to a

2

more local search as the number of function evaluations approaches the maximum allowable

3

limit. The adjustment from global to local search is achieved by dynamically and probabilistically

4

reducing the number of dimensions in the neighborhood (i.e. the set of decision variables

5

modified from their best value). Candidate solutions are sampled from the neighborhood by

6

perturbing only the randomly selected decision variables from the current solution. These

7

perturbation magnitudes are randomly sampled from a normal distribution with a mean of zero

8

for the continuous version of DDS. These features of the DDS algorithm ensure that it is as

9

simple and parsimonious as possible. DDS is a greedy type of algorithm since the current

10

solution, also the best solution identified so far, is never updated with a solution that has an

11

inferior value of the objective function. The algorithm is unique compared with current

12

optimization algorithms because of the way the neighborhood is dynamically adjusted by

13

changing the dimension of the search. The DDS perturbation variances remain constant and the

14

number of decision variables perturbed from their current best value decreases as the number of

15

function evaluations approaches the maximum function evaluation limit. This key feature of

16

DDS was motivated by experience with manual calibration of watershed models where early in

17

the calibration exercise relatively poor solutions suggested the simultaneous modification of a

18

number of decision variables but as the calibration results improved, it became necessary to only

19

modify one or perhaps a few decision variables simultaneously so that the current gain in

20

calibration results was not lost.

21

The discrete DDS algorithm pseudocode is given in Figure 1, and the changes relative to the

22

original DDS algorithm are in Steps 3 and 5. The only user-defined algorithm parameter is the

23

scalar neighborhood size perturbation parameter (r), which defines the standard deviation of the

24

random perturbation size as a fraction of the decision variable range. In the discrete DDS

25

algorithm, the decision variables are integers from 1 to the number of discrete options for each

26

decision variable (ximax). The objective function must translate or map these option numbers to

27

discrete pipe diameters in the WDS optimization case. Consistent with the continuous version of 7

1

DDS, a default value of the r parameter for discrete DDS is recommended as 0.2 (and used in this

2

study). In the original DDS algorithm, the perturbation magnitude for each decision variable is

3

sampled from a normal probability distribution. In Step 3 of discrete DDS (Figure 1), the

4

perturbation magnitude is randomly sampled from a discrete probability distribution that

5

approximates a normal distribution (see example probability mass functions in Figure 2). In the

6

continuous and discrete DDS algorithms, r = 0.2 yields a sampling range that practically spans

7

the normalized decision variable range for a current decision variable value halfway between the

8

decision variable bounds. This sampling region size is designed to allow the algorithm to escape

9

regions around poor local minima. The r parameter is a scaling factor assigning the same relative

10

variation to each decision variable (relative to the decision variable range).

11

The one-dimensional decision variable perturbations in Step 3 of the DDS algorithm can

12

generate new decision variable values outside of the decision variable bounds (e.g. 1 and ximax).

13

In order to ensure that each one-dimensional perturbation results in a new decision variable that

14

respects the bounds, the DDS and discrete DDS algorithms define reflecting boundaries (See Step

15

3 of Figure 1). In the discrete DDS algorithm, the candidate decision variable values are initially

16

treated as continuous random variables and the reflecting boundaries for the decision variables

17

within the algorithm are defined to be 0.5 and 0.5 + ximax. Once a candidate decision variable

18

value within these reflecting boundaries is sampled, it is rounded to the nearest integer value to

19

represent the discrete option number (e.g. 1, 2, 3, … , ximax). This reflecting boundary approach

20

allows decision variable values to more easily approach their minimum or maximum values in

21

comparison with a simple perturbation resampling approach (truncated normal distribution) for

22

ensuring decision variable boundaries are respected. In the event that a perturbed decision

23

variable has a candidate discrete option number that is the same as the option number of the

24

current best solution (i.e. the decision variable is not perturbed), a new discrete option number is

25

sampled from a discrete uniform probability distribution. The resultant probability mass function

26

for the decision variable option number is given for four examples in Figure 2 based on Step 3 of

27

Figure 1. 8

1

The maximum number of objective function evaluations (M) is an algorithm input (like the

2

initial solution) rather than an algorithm parameter because it should be set according to the

3

problem specific available (or desired) computational time [see Gibbs et al., 2008]. The value of

4

M therefore depends on the time to compute the objective function and the available

5

computational resources or time. Except for the most trivial objective functions, essentially 100%

6

of discrete DDS execution time is associated with the objective function evaluation. Recall that

7

discrete DDS scales the search strategy from global in the initial stages of the search to more

8

local in the final stages of the search regardless of whether M is 1000 or 1,000,000 function

9

evaluations. In the absence of a specific initial solution, discrete DDS is initialized to the best of a

10

small number (maximum of 0.005M and 5) of uniformly sampled random solutions.

11

In the continuous DDS algorithm, depending on the problem dimension and maximum

12

objective function evaluation limit (M), a significant proportion of function evaluations in the

13

latter half of the search would evaluate a change in only one decision variable relative to the

14

current best solution. Given the general constrained WDS problem, we know a priori that the

15

feasible high quality solutions discrete DDS (or any other algorithm) identifies by the later stages

16

of the search will be very close to infeasibility. We also know that the only way to minimize the

17

objective function we define for discrete DDS is to select a smaller pipe diameter and thus reduce

18

nodal pressures somewhere in the network. Being close to infeasibility means that the original

19

DDS perturbation strategy that happens late in the search (changing only one decision variable)

20

will quickly become futile and all possible single pipe diameter reductions will result in

21

infeasible solutions. Therefore, a new primary stopping condition was added to discrete DDS in

22

Step 5 to stop the search before all M function evaluations are utilized (and thus save the

23

remaining computation time for more productive search strategies) based on the probability of

24

decision variable perturbation, P(i).

25

The number of dimensions selected for perturbation at each discrete DDS iteration, as

26

determined by the first two bullets of Step 2 in Figure 1, follows a binomial probability

27

distribution parameterized by P(i). It is desired that more than one dimension is selected for 9

1

perturbation at each iteration, which implies that an appropriate termination point is when the

2

expected value of the binomial random variable is one. The corresponding probability value at

3

this expected value is 1/D (D is the number of decision variables) leading to a termination

4

criterion of P(i) < 1/D.

5

For the moderately sized WDS case studies solved (21-42 dimensions) with computational

6

budgets ranging from 10,000 to 1,000,000 objective function evaluations, this new stopping

7

condition terminates discrete DDS after 52% to 80% of M allowable function evaluations are

8

conducted. In networks with an order of magnitude of more decision variables (more than 200),

9

for any computational budget fewer than 10,000,000 function evaluations, discrete DDS will only

10

stop after 90% or more of M allowable function evaluations are utilized. The complete HD-DDS

11

algorithm and its other component search strategies (sections 2.1.3, 2.1.4, and 2.1.5) define how

12

the remaining computational budget is utilized after an initial optimization with discrete DDS

13

terminates.

14

A preliminary version of discrete DDS was first defined in Tolson et al. [2008]. However,

15

discrete DDS in Figure 1 of this study differs from Tolson et al. [2008] in two ways. First, we use

16

a slightly modified neighborhood definition more consistent with the original DDS algorithm in

17

Tolson and Shoemaker [2007]. Another distinction is that Tolson et al. [2008] allowed the

18

preliminary discrete DDS to utilize all M function evaluations rather than stopping the algorithm

19

early.

20

2.1.2. Constraint-Handling with Discrete DDS

21

Discrete DDS is not a population based algorithm and only one solution (the best found so

22

far) influences the region of decision variable space that is sampled. As shown in Step 5 of Figure

23

1, discrete DDS (like DDS) is not impacted by objective function scaling. Only the relative ranks

24

of the best solution found so far and the candidate solution determine whether the best solution

25

should be updated. Any update to the best solution moves it to a different point in decision space,

26

around which future candidate solutions are centered. This aspect of DDS and discrete DDS 10

1

makes handling constraints very straightforward without the need for penalty function

2

parameters. The constraint handling approach described below is equivalent to the method

3

described in Deb [2000] for constraint handling in GAs using tournament selection (where

4

objective function scaling also has no impact on the optimization). The key to the approach in

5

Deb [2000] is that the objective function is defined such that any infeasible solution always has

6

an objective function value that is worse than the objective function value of any feasible

7

solution. In a WDS design problem where costs are to be minimized subject to minimum pressure

8

requirements across the network, the worst cost feasible solution is known before any

9

optimization as the cost of the solution specifying all pipes at their maximum diameter. The only

10

other requirement for the constraint handling technique in Deb [2000] is to quantify the relative

11

magnitude of constraint violations for infeasible solutions so that the relative quality of two

12

infeasible solutions can be compared. The steps to evaluate the objective function in discrete

13

DDS are outlined in Figure 3.

14

The evaluation of the objective function in Figure 3 has a penalty term, but there is no need to

15

scale it to a different magnitude with a case study specific penalty multiplier and/or exponent.

16

Instead, the objective function definition implements the following logic:

17



18 19

assigned a better objective function value. 

20 21 22

Between two infeasible solutions, the one with the least total pressure violation is always

Between an infeasible and a feasible solution, the feasible one is always assigned a better objective function value.



Between two feasible solutions, the one which is less costly is always assigned a better objective function value.

23

In addition to the benefit of requiring no case study specific penalty parameter values

24

(which generally require experimentation to determine), our definition of the objective function

25

yields the significant benefit of eliminating the need to evaluate the hydraulics of a large number

26

of solutions generated by discrete DDS. The multi-objective optimization study by Prasad and

11

1

Park [2004] is the only other WDS study that we are aware of to utilize the constraint handling

2

strategy in Deb [2000].

3

2.1.3. One-pipe Change Local Search (L1)

4

This local search technique starts at a feasible solution and cycles through all the possible

5

ways to perturb the current best solution by reducing the diameter of one pipe at a time (thus

6

saving costs). Each perturbed solution is evaluated for cost and feasibility. The L1 search

7

continues until it has evaluated a maximum number of objective function evaluations or until it

8

has enumerated all possible one-pipe changes without finding any improvement to the current

9

best solution and therefore confirms that a local minimum has been located. The solution is a

10

local minimum in that no better solution exists that differs from the current best solution in only

11

one decision variable (pipe diameter). Pseudocode for implementing L1 will be archived with this

12

paper online1.

13

Note that L1 is implemented such that it is always evaluating one-pipe changes relative to

14

the current best known solution. In other words, whenever L1 finds a new best solution, the

15

current best solution is updated and L1 searches in the vicinity of the updated current best

16

solution. Therefore, the order of enumeration (starting at pipe D instead of pipe 1) can change

17

the final best solution returned by L1.

18

2.1.4. Two-pipe Change Local Search (L2)

19

This local search technique starts at a feasible solution and cycles through all the possible

20

ways to perturb the current best solution by only two pipes where one pipe has the diameter

21

increased and the other pipe has the diameter decreased. In this study, the L2 search is only

22

initiated at feasible solutions that are already confirmed to be a local minimum with respect to L1.

23

As Gessler [1985] suggested in his pipe design enumeration strategy, solutions that have a higher

1

Auxiliary material is available at ftp://ftp.agu.org/***/L1.eps.

12

1

cost than the current best solution are not evaluated for their hydraulics in the L2 search. The L2

2

search continues until it has evaluated a maximum number of objective function evaluations, or it

3

has enumerated all possible two-pipe changes without finding any improvement to the current

4

best solution and therefore confirms that a local minimum has been located. The solution is a

5

local minimum in that no better solution exists that differs from the current best solution in only

6

two decision variables (pipe diameters). Pseudocode for implementing L2 will be archived with

7

this paper online1.

8

Note that L2 is implemented such that it enumerates all possible two-pipe changes from the

9

current best solution to identify the best two-pipe change. This is a different enumeration

10

approach in comparison to L1. Only after all two-pipe change possibilities are enumerated in L2 is

11

the current best solution updated and the enumeration process repeated. Therefore, this local

12

search method is defined so that different orders of enumeration do not affect the results if L2

13

converges to a local minimum.

14

In L2, the solution being enumerated only has the objective function evaluated if it is less

15

expensive than the current best L2 solution. This is quickly determined based only on comparing

16

the costs associated with the two decision variables (pipe diameters) being changed relative to the

17

corresponding costs in the current best L2 solution.

18

2.1.5. HD-DDS Algorithm Definition

19

The explicit design goals that guided our development of the Hybrid Discrete Dynamically

20

Dimensioned Search (HD-DDS) Algorithm were, in highest to lowest priority, to develop an

21

algorithm that 1) reliably returned good quality approximations of the global optimum 2) was

22

capable of solving WDN design problems to optimality and 3) could achieve these goals in a way

23

that was computationally efficient. In addition, the algorithm was designed to be parsimonious in

24

the number of algorithm parameters such that achieving the above goals did not necessarily

1

Auxiliary material is available at ftp://ftp.agu.org/***/L2.eps.

13

1

require parameter tuning experiments for each case study. The definition of a good quality

2

approximation is clearly specific to each case study and even the computational budget used to

3

solve the problem, but in general, is taken to be a solution whose cost is small enough to satisfy

4

the designer such that further optimization would be deemed unnecessary.

5

characteristic that the algorithm inherits from the continuous DDS algorithm is the ability to

6

effectively scale to various computational budgets such that it is capable of generating good

7 8 9 10 11 12 13 14 15

results without algorithm parameter tuning. The HD-DDS algorithm component is designed to overcome the shortcomings of other individual component search strategies. First and foremost, as described in section 2.1.1, the discrete DDS search is predictably unproductive with the original DDS algorithm stopping criterion, and thus HD-DDS invokes alternative search strategies after discrete DDS reaches this unproductive stage (when P(i) < 1/D). Unlike other hybridized algorithms (such as a ACO + a local search), the point at which to terminate a global search in favour of a new search strategy in HD-DDS is known a priori and thus does not require the definition of various algorithm specific convergence measures. The component search combination strategy defining HD-DDS is given in

16 17

Figure 4. Initially (Step 0 in

A design

18 19

Figure 4), HD-DDS requires all problem inputs to be specified. This includes the user-

20

defined computational budget, Mtotal, which can be expended in solving the problem. Mtotal is a

21

quantity that users should determine based on their problem specific available computation time

22

and the average computation time required for an objective function evaluation that evaluates

23

both cost and hydraulics. As discussed later in section 4, because HD-DDS can skip the

24

hydraulics evaluation for a large number of candidate solutions, HD-DDS will terminate much

25

quicker than what would be estimated in this way. The only algorithm parameter in HD-DDS is

26

the discrete DDS parameter r, which has a default of 0.2 that is utilized for all HD-DDS analyses

27 28

presented in this paper. The first search strategy implemented in Step 1 of

29 30

Figure 4 is a global search with discrete DDS for a maximum of M objective function

31

evaluations, where M in HD-DDS is a variable tracking the remaining available number of 14

1

objective function evaluations. The second search strategy in Step 2 of HD-DDS is defined to be

2

the L1 local search (one-pipe change), which polishes the discrete DDS solution to a local

3

minimum with respect to L1. Results show this enumerative search is typically very quick (often

4 5

less than D evaluations of EPANET2) when polishing a discrete DDS solution. Step 3 in HD-DDS (

6 7 8 9 10 11

Figure 4) is defined as a second independent discrete DDS search from a new initial solution followed by another L1 search to polish the second discrete DDS solution. This second discrete DDS search has a smaller M than the first and therefore will utilize fewer objective function evaluations. At the end of Step 3, a second potentially new solution that is a local minimum with respect to L1 has been located, which is referred to as xBbest. Step 4 in HD-DDS (

12 13

Figure 4) implements the first L2 local search to polish the best of the two HD-DDS

14

solutions found so far. With very large networks, L2 can be extremely slow to converge and

15

confirm that a local minimum has been reached.

16

reasonably good quality solution, L2 typically converges in a few thousand objective function

17

evaluations in HD-DDS for case studies with up to 42 decision variables. Step 5 in HD-DDS

18

implements the second L2 local search to polish the other HD-DDS solution that is confirmed to

19

be a local minimum with respect to L1.

However, as results will show, with a

20

HD-DDS can terminate when either the total computational budget (Mtotal) is exceeded or

21

when all five steps are completed in fewer than Mtotal objective function evaluations. For a fixed

22

and reasonably large computational budget (e.g. 100,000), as the network size decreases (number

23

of decision variables decrease), the likelihood that HD-DDS will be able to complete all five

24

steps increases. However, as network size increases, it becomes more and more likely that HD-

25

DDS will terminate without a second discrete DDS search. In other words, HD-DDS reduces to

26

only Steps 1 and 2 when the number of decision variables becomes extremely large because the

27

user defined computation limit will be exceeded before Step 2 finishes. This behavior is based on

28

the assumption that for a fixed and practical computational budget, the globally optimal solution

15

1

to an extremely large problem is virtually impossible to find, and the best approach would be to

2

conduct the longest possible global search.

3

HD-DDS was designed to always perform a second global search (Step 3) before the first

4

L2 local search polishing step. The main purpose of a second global search is to act as a safety

5

mechanism to guard against a very poor solution from the first discrete DDS trial. To maximize

6

the chances the safety mechanism works, this second global search requires the largest possible

7

computation budget and therefore must be executed prior to L2 (which can take an indeterminate

8

amount of time). Furthermore, our belief in general is that the best chance to significantly

9

improve upon a particularly poor solution is to conduct a new global search rather than polish the

10

particularly poor solution. This design choice has no impact on final solution quality in cases

11

where the computation budget is large enough and/or the problem dimension is small enough to

12

allow all five steps in HD-DDS terminate completely in fewer than Mtotal objective function

13

evaluations. There are some situations (particularly with problems having hundreds of decision

14

variables) where the choice to conduct a second global search (followed by L1) uses all or most of

15

the remaining computational budget and thus precludes or reduces computation time available for

16

the L2 local search. In these situations with such large networks, the L2 local search can require

17

an incredible number of objective function evaluations to enumerate all possible two-pipe

18

changes such that L2 terminates before it confirms the solution is a local minimum. Despite this,

19

the optimal design choice between L2 or second global search is certainly case study dependent.

20

Results will show that our design choice has very little impact on HD-DDS results.

21

2.2. Benchmark Optimization Algorithms

22

The main focus of this study is the introduction of the HD-DDS algorithm and its

23

performance comparison with alternative algorithms. The design case studies and benchmark

24

algorithms were selected from the literature based on whether the WDS case study could be

25

replicated exactly, and whether the algorithm results were presented for multiple optimization

26

trials. Exact replication of a WDS case study meant that comparative algorithm results must have 16

1

been generated with EPANET2 using metric units. Most WDS optimization algorithms are

2

stochastic in that they can and often do generate a different final solution for a different random

3

seed and/or different initial solution. Therefore, objectively assessing relative algorithm

4

performance requires multiple independent optimization trials.

5

Three recent studies that meet the above criteria and thus provide the majority of the

6

comparative algorithm results used in this study are Zecchin et al. [2007], Reca and Martinez

7

[2006], and Montalvo et al. [2008]. Zecchin et al. [2007] test the performance of five different

8

Ant Colony Optimization (ACO) algorithms on some standard WDS case studies. They report

9

that the MMAS ACO algorithm outperforms all other algorithms applied to their case studies in

10

the literature. Reca and Martinez [2006] apply a GA called GENOME to a much larger scale

11

WDS optimization problem. Montalvo et al. [2008] introduce a Particle Swarm Optimization

12

(PSO) algorithm variant to some standard WDS case studies and show that their PSO variant

13

outperforms a standard discrete PSO algorithm. Details of the WDS case studies selected for HD-

14

DDS testing from the above studies are given in the following section.

15

2.3. Benchmark WDS Design Studies

16

The first three example WDS design problems in this study, the New York Tunnels, Doubled

17

New York Tunnels, and Hanoi problems, are equivalent to those used in Zecchin et al. [2007].

18

For complete details of these three case studies, readers are referred to Zecchin et al. [2005]. The

19

GoYang WDS problem was introduced by Kim et al. [1994] and is also utilized as our fourth case

20

study. The fifth and final case study is the Balerma WDS problem, which was first proposed by

21

Reca and Martinez [2006]. The next few paragraphs provide an overview of each WDS case

22

study, and readers should consult the references given for case study details (e.g. discrete pipe

23

diameter options, pipe length, nodal head requirements etc.). Note that all EPANET2 units are

24

metric in each of the case studies.

25

The New York Tunnels Problem (NYTP), originally considered in Schaake and Lai [1969],

26

involves the rehabilitation of an existing WDS by specifying design options for the 21 tunnels in 17

1

the system. There are 16 design options for each tunnel (parallelization with one of 15 tunnel

2

sizes or a do-nothing option), thus defining a search space size of 1621 (approximately

3

1.93×1025). The Doubled New York Tunnels problem (NYTP2) was originally proposed by

4

Zecchin et al. [2005]. This network has 42 pipes to be sized, and each has 16 options. This

5

defines a search space of 1642 (approximately 3.74×1050) for NYTP2.

6

The Hanoi Problem (HP) was first reported on in Fujiwara and Khang [1990] and has a

7

larger search space size (2.87x1026 solutions) than the NYTP based on 32 pipes to be sized and 6

8

discrete pipe diameters. HP problem details are available in Wu et al. [2001]. The GoYang

9

problem (GYP) represents a South Korean WDS and was first proposed by Kim et al. [1994] and

10

has 30 pipes to be sized with 8 diameter options per pipe to define a search space size of 830

11

(approximately 1.24×1027).

12

The Balerma irrigation network problem (BP) from Reca and Martinez [2006] is a large and

13

complex network that has 443 demand nodes, 454 pipes, 8 loops, and 4 reservoirs. Each of the

14

454 pipes to be sized has ten possible diameters and thus defines a search space size of 10454.

15

Based on a review of the current literature, the best current known feasible solutions (using

16

EPANET2) are $38.638 million for the NYTP [Maier et al., 2003], $77.276 million for the

17

NYTP2 [Zecchin et al., 2005], $6.081 million for the HP [Perelman and Ostfeld 2007], and

18

€2.302 million for the BP [Reca and Martinez 2006].

19

algorithm results for GYP with EPANET2 as the hydraulic solver are unavailable and thus HD-

20

DDS is not compared to any other algorithm for the GYP.

21

2.4. Optimization Model Formulation for HD-DDS

22

Currently, published optimization

The HD-DDS algorithm solves the following optimization model for each WDS case study:

23

Min F( x ) ,

24 25

x

s.t.



(1)



xd  1, 2, , xdmax ,  d = 1, …, D

18

1

where F(x) is the objective function value for decision vector x = [x1 … xD] as determined by the

2

procedure in Figure 3, and xd is the integer-valued pipe diameter option number for pipe d and is

3

between option 1 (the smallest diameter) and the maximum diameter option, x dmax , for all D pipes

4

in the network to be sized. The minimum required heads for the nodes in the network (e.g. himin

5

for node i in Figure 3) and nonlinear cost equations vary with each case study and must be

6

specified (readers can find these case study specific values in the case study references noted in

7

section 2.3).

8

2.5. Outline of Algorithm Comparisons

9

The performance of different algorithms is generally compared for a similar computational

10

budget as measured by the total number of objective function evaluations utilized in the

11

optimization run. The computational budget we report for all algorithms is measured as the

12

budget required for the algorithms to terminate. Although this is not the measure that is reported

13

in Zecchin et al. [2007] for the MMAS ACO algorithm, the measure used in our study reflects the

14

fact that in any new problem, with an unknown best solution, the user will not generally have any

15

knowledge or reason to stop the algorithm before it terminates normally.

16

Due to the stochastic nature of most heuristic global optimization algorithms, their relative

17

performance must be assessed over multiple independent optimization trials, each initialized to

18

independent populations or solutions. Previously published algorithm results are compared to

19

HD-DDS results using 10 to 50 optimization trials. Optimization algorithm comparisons cover

20

21-, 30-, 32-, 42- and 454-dimensional (i.e. the number of decision variables or pipes to be sized)

21

problems, and the size of the search spaces ranges from 1.93x1025 to 1.0x10454 possible solutions.

22

The maximum number of objective function evaluations, Mtotal, per HD-DDS optimization trial

23

varies from 10,000 to 10,000,000. Given that average algorithm performance across multiple

24

independent optimization trials does not provide a complete picture of results, the distribution or

25

range of the HD-DDS and benchmark algorithm solutions are also assessed.

19

1

The initial solution for HD-DDS is generated as described in section 2.1.1, and the

2

neighborhood size parameter, r, is set to the default value of 0.2 for all HD-DDS trials (no

3

parameter setting experimentation was performed on r). In some of the case studies, L1 or L2

4

local searches initiated in HD-DDS before the remaining available computational budget is

5

exceeded were typically allowed to terminate at a local minimum even if that meant exceeding

6

Mtotal by a few thousand objective function evaluations on average. In contrast, comparative

7

MMAS ACO algorithm results from Zecchin et al. [2007] are based on MMAS ACO algorithm

8

parameters that were optimized independently for each case study using millions of EPANET2

9

simulations. Similarly, the comparative results for the GENOME GA and PSO variant algorithms

10

were also derived from some experimental optimization runs to identify good algorithm

11

parameters.

12 13

3.

Results

14 15

The results are presented here in two sections. Section 3.1 assesses the importance and effect

16

of each step of the HD-DDS algorithm for four case studies. Comparative algorithm performance

17

results are presented in section 3.2

18

3.1. HD-DDS Component Performance Assessment

19

For all HD-DDS optimization trials, at the end of each HD-DDS step in

20 21

Figure 4, the best HD-DDS solution found so far and the corresponding number of objective

22

function evaluations are recorded. This information is shown in Figure 5 for the NYTP, NYTP2,

23

HP, and GYP benchmarks. In addition to presenting results from all individual optimization

24

trials, the average cost of the best solution found so far is shown in Figure 5 where objective

25

function values are plotted against the average number of function evaluations for clarity because

26

for Steps 2, 3, 4 and 5, the number of utilized function evaluations varies between optimization 20

1

trials. The difference between each step on the x-axis in Figure 5 shows the average

2

computational requirements of each Step in HD-DDS.

3

For each case study in Figure 5, Step 2 in HD-DDS (the L1, one-pipe change local search)

4

requires a negligible number (~10) of objective function evaluations to determine that discrete

5

DDS terminated at a local minimum with respect to L1. In fact, in all HD-DDS optimization

6

trials, L1 never improved upon a discrete DDS solution, indicating that discrete DDS always

7

terminated at a local minimum with respect to L1. With a much smaller computational budget

8

and/or much larger network (e.g. see BP results for Mtotal ≤ 10,000 in section 3.2.1), discrete DDS

9

will not always terminate at a local minimum with respect to L1, and thus L1 can improve upon

10

discrete DDS solutions.

11

The computational budget for the second discrete DDS search (Step 3) is significantly

12

smaller than that for the first discrete DDS search for all four case study results in Figure 5 (21%

13

- 41% of the budget of the first discrete DDS search). Despite the substantial decrease in

14

computational budget, results of Steps 2 and 3 in Figure 5 show that this second, quicker discrete

15

DDS search does sometimes improve upon the first DDS search result (for all four case studies,

16

Step 3 improves best solution in 20% to 32% of optimization trials). Most notable in Figure 5 is

17

that the second DDS search operates to eliminate the worst cost solutions from Step 2 in all four

18

case studies.

19

Results in Figure 5, at each Step of HD-DDS, also present a count of the number of

20

optimization trials that have located the best known solution for each of the case studies. Discrete

21

DDS (Steps 1 and 3 of HD-DDS) generally does not find the best known solution (except for a

22

small number of trials in NYTP and GYP). Results for Steps 4 and 5 of HD-DDS in Figure 5

23

clearly show that the L2 (two-pipe change) local search is capable of polishing discrete DDS

24

solutions (from Steps 1 and 2) and returning the best known solution (with 85% frequency for

25

NYTP and NYTP2). The average computational burden and percentage of optimization trials

26

with improved solutions due to L2 in Step 4 (which terminates with a local minimum with respect

27

to L2) was 1421, 9031, 3250, and 586 function evaluations and 90%, 56%, 100%, and 66% for 21

1

NYTP, NYTP2, HP, and GYP case studies, respectively. The average computational burden for

2

Step 5 (second L2 search which terminates with a local minimum with respect to L2) increases to

3

2302, 15403, 4487, and 649 function evaluations on average for the NYTP, NYTP2, HP, and

4

GYP case studies, respectively, because the second L2 search is typically starting from a lower

5

quality discrete DDS solution. Results for NYTP and NYTP2 in Figure 5 demonstrate that Step 5

6

can improve results indicating that a second L2 search starting from a lower quality initial

7

solution can be fruitful. The percentage of optimization trials with improved solutions due to L2

8

in Step 5 is 32%, 25%, 14%, and 6% for the NYTP, NYTP2, HP, and GYP case studies,

9

respectively.

10

3.2. HD-DDS Performance Relative to Benchmark Algorithms

11

The first comparative results are for HD-DDS and the MMAS ACO algorithm results in

12

Zecchin et al. [2007] for the NYTP, HP and NYTP2. Figure 6 shows the empirical cumulative

13

distribution function (CDF) for HD-DDS and MMAS ACO of the final best objective function

14

values (costs) from all optimization trials. HD-DDS results stochastically dominate MMAS ACO

15

results in all three case studies because for any desirable cost target identified by a decision-

16

maker, HD-DDS always has an equal or higher probability of satisfying the cost target than

17

MMAS ACO. The near vertical lines for NYTP and NYTP2 indicate that HD-DDS yields the

18

best known solution with a high reliability (more than 80%).

19

optimization trials were better than the best MMAS ACO solution. HD-DDS avoided the worst

20

solutions identified by MMAS ACO. The superior performance of HD-DDS over MMAS ACO

21

is even more noteworthy considering there was no algorithm parameter experimentation with

22

HD-DDS, and there were extensive case study specific parameter setting experiments conducted

23

for MMAS ACO, as discussed previously.

24

3.2.1. Large Scale WDS: Balerma Case Study

25

The Balerma network has 454 pipes to be sized (D=454), and as a result, Step 1 in HD-DDS (see

For HP, eight HD-DDS

22

1 2

Figure 4) uses more than 96% of any of the computational budgets specified in our study

3

(1,000 ≤ Mtotal ≤ 10,000,000). Note that Step 1 of HD-DDS (the first discrete DDS search)

4

terminates only when P(i), as calculated in Step 2 of discrete DDS (see Figure 1), is less than 1/D.

5

The remaining 4% or less of the computational budget is utilized by Step 2 (L1 search) and then

6

any remaining budget is dedicated to Step 3 of HD-DDS (second global search followed by

7

another typically incomplete L1 search).

8

Figure 7 shows all of the HD-DDS results generated for Balerma as the number of

9

objective function evaluations are increased from 1000 to 10,000,000. All HD-DDS results are

10

based on 10 optimization trials. As expected, Figure 7 shows that HD-DDS performance

11

improves with a larger computational budget. It is important to recall that HD-DDS scales the

12

search to the user input Mtotal, and thus separate independent optimization trials are used to

13

generate results for different computational budgets. For example, the best solution after 10,000

14

objective function evaluations in HD-DDS with Mtotal=100,000 will not be equal to the final best

15

solution in HD-DDS with Mtotal=10,000.

16

Comparative algorithm performance is also shown in Figure 7 using results for MSATS

17

and GENOME GA reported in Reca et al. [2008] and GENOME GA results reported in Reca and

18

Martinez [2006]. For the same computational budget (10,000,000 objective function evaluations),

19

HD-DDS clearly outperforms the GENOME GA, as the worst HD-DDS solution is better than

20

the best GENOME GA solution by nearly 200,000 Euros. Even with 1/100th of the computational

21

budget (Mtotal=100,000), HD-DDS still outperforms the GENOME GA as the worst HD-DDS

22

result costs nearly 100,000 Euros less than the best GENOME GA result after 10,000,000

23

function evaluations.

24

In the interest of determining the best possible solution, the best HD-DDS solution with

25

Mtotal=10,000,000 (cost of €1,956,226) was passed onto the L2 local search to polish the solution.

26

After an additional 20 million EPANET2 objective function evaluations in L2, the minimum cost

23

1

solution improved to €1,940,923 (also shown in Figure 7). Note that L2 was manually terminated

2

prior to confirmation a local solution was identified.

3

The HD-DDS results for Mtotal=10,000 and Mtotal=1000 in Figure 7 demonstrate that even

4

with a severely restricted computational budget, HD-DDS can generate reasonable solutions, all

5

of which are feasible. These HD-DDS results are much better than the GENOME GA and

6

MSATS results from Reca et al. [2008] after 45,000 objective function evaluations. Note that the

7

HD-DDS results for Mtotal=1,000 utilized an average of 2,900 objective function evaluations

8

rather than 1000 because the L1 one-pipe change search was only terminated when it returned a

9

local solution with respect to L1 (which required approximately 1900 additional function

10

evaluations). In fact, the L1 search drastically improved the average discrete DDS solution quality

11

from 5.021 to 3.080 million Euros. Application of only L1 without the preliminary discrete DDS

12

search was ineffective in large part because without the discrete DDS solution, L1 could not

13

quickly identify a feasible solution.

14

Some final tests were performed to compare HD-DDS performance if L2 was performed

15

instead of the second global search. Although performing L2 yields reliable but very small

16

improvements (0.1%-0.2% on average) in the best cost solution after Step 2 of HD-DDS in

17

Figure 4, performing the second global search is capable of yielding significantly larger best cost

18

improvements (more than 2%) much less frequently. Therefore, as designed the algorithm

19

forgoes very small improvements under L2 for the significantly higher but less frequent

20

improvements capable with second global search.

21

3.2.2. HD-DDS Performance Comparison Summary

22

Table 2 summarizes and compares the results of HD-DDS and other algorithms previously

23

noted in sections 3.2 and 3.2.1. The main difference in addition to compressing all previous

24

graphical results into one table is that the solution quality is also measured with respect to the

25

percent deviation from the best known solution. Table 2 also includes new results for a few other

26

algorithms and HD-DDS computational budgets. Results for other algorithms in Table 2 are only 24

1

included where it was possible to confirm that the algorithms were applied to the exact same

2

optimization formulation (e.g. EPANET2 was used with metric units).

3

Results in Table 2 pair each HD-DDS result with another comparative algorithm result,

4

where HD-DDS is typically applied with a similar number of objection function evaluations, and

5

demonstrate the excellent overall performance of HD-DDS. In all algorithm comparisons, the

6

median best costs found by HD-DDS are always equal to or lower than the costs obtained using

7

the other algorithms. Importantly, the maximum cost solutions found by HD-DDS are always

8

better than those found by the comparative algorithms for all case studies (HD-DDS is 1.2% to

9

21.3% closer to the best known solution). For example, the worst HD-DDS solution for NYTP2

10

is nearly 2 million dollars less than the worst MMAS ACO solution and the worst HD-DDS

11

solution for NYTP is just over 8 million dollars less than the worst PSO variant solution. The

12

minimum best costs found by HD-DDS are equal to or lower than the costs obtained using all

13

other algorithms in all comparisons. With the exception of the 150,000 x 10 results for HP, Table

14

2 shows HD-DDS always returns the best known solution with a higher frequency than other

15

algorithms. For example, HD-DDS finds the best known NYTP solution in 86% of optimization

16

trials compared to 30% for the PSO variant despite HD-DDS using 30,000 fewer objective

17

function evaluations.

18

The best HD-DDS solutions for NYTP and NYTP2 are the same as the best known

19

solutions found in Zecchin et al. [2007]. The best HD-DDS solution for HP is the same as the

20

best known solution as reported in Perelman and Ostfeld [2007]. The best HD-DDS solution for

21

BP is a new best known solution. All of these HD-DDS identified best known solutions will be

22

archived with this paper online1.

23 24

4.

Discussion

25

1

Auxiliary material is available at ftp://ftp.agu.org/***/Archive_best_solutions.txt.

25

1

The default value of the neighborhood perturbation size parameter, r, of 0.2 produced

2

excellent results compared to all other algorithms for the WDS case studies reported on in section

3

3. These good results cover 21- to 454-dimensional problems and are based on computational

4

budgets ranging from 1,000 to 10,000,000 objective function evaluations. Therefore, the default

5

value for r appears robust and is suggested for future HD-DDS applications. Results also showed

6

that each search strategy (or step) in HD-DDS played an important role in at least one case study.

7

HD-DDS has a very large computational efficiency advantage over most other WDS

8

optimization algorithms that is not obvious from results reported in section 3. For consistency

9

with previous studies, the computational budget of HD-DDS was defined with respect to the total

10

number of objective function evaluations. However, the objective function evaluation strategy

11

with constraint handling (see Figure 3) and the computationally efficient implementation of L2

12

enable HD-DDS to evaluate the solution quality to a sufficient level without simulating the

13

hydraulics of the solution (e.g. an EPANET2 simulation). Therefore, even though, for example,

14

HD-DDS utilized approximately 46,000 objective function evaluations to optimize the NYTP

15

with Mtotal=50,000, EPANET2 simulations were not required for approximately 33,000 or 72% of

16

all HD-DDS solutions evaluated. Based on actual run times for HD-DDS, results show that our

17

HD-DDS algorithm as implemented runs 50% faster than it would have if EPANET2 was used to

18

evaluate all 46,000 solutions identified in HD-DDS. Balerma results for Mtotal=100,000 were

19

similar in that EPANET2 simulations were not required for approximately 71% of all HD-DDS

20

solutions evaluated. Based on HD-DDS run times for Balerma, results show that our HD-DDS

21

algorithm as implemented runs 67% faster than it would have if EPANET2 was used to evaluate

22

all solutions identified in HD-DDS. The relative computational efficiency gain of HD-DDS over

23

other optimization algorithms that must evaluate hydraulics of every candidate solution becomes

24

larger as the computational demand of the hydraulics simulation increases. If the EPANET2

25

evaluations accounted for nearly 100% of HD-DDS computation time then HD-DDS could be

26

more than 70% faster than other WDS optimization algorithms like MMAS ACO [Zecchin et al.,

26

1

2007] and the roulette wheel selection based GENOME GA [Reca and Martinez, 2006] in

2

evaluating the same number of objective functions.

3

To our knowledge, this paper was the first time the constraint handling strategy in Deb

4

[2000] was used in a single objective WDS optimization application, and overall, the excellent

5

results of HD-DDS suggest the strategy works very well. Every discrete DDS optimization trial

6

we performed always returned a final solution that was feasible, even in the HP for which

7

multiple studies report difficulty in locating any feasible solution due to its small feasible search

8

space [Euseff and Lansey, 2003; Zecchin et al., 2005 and Zecchin et al., 2007]. In order to further

9

demonstrate the excellent performance of HD-DDS with our constraint handling approach for

10

HP, 10 independent trials initialized at the most infeasible solution (all pipes at their minimum

11

diameter) with Mtotal=10,000 were conducted. Each of these runs were terminated at exactly

12

10,000 function evaluations, and all returned feasible solutions with an average cost of $6.299

13

million and a worst cost of $6.375 million.

14

For a fixed computational budget and specific case study, it is possible that there are better

15

ways to combine discrete DDS, one-pipe and two-pipe change algorithms than the HD-DDS

16

algorithm presented. In other words, we do not claim that HD-DDS is the optimal way to

17

configure these three search strategies across all case studies and all computational budgets. The

18

optimal configuration of these three search strategies is almost certainly case study and

19

computational budget specific. Instead, we have demonstrated a very robust and parsimonious

20

way to combine these strategies. HD-DDS performs two independent global searches before

21

spending an unknown amount of effort to polish the best solution with the two-pipe change local

22

search. Our results show that relative to available benchmark algorithm performance, for similar

23

computational budgets, HD-DDS performs equivalent to or better than all algorithms in our

24

comparison in the five case studies considered here. Therefore, experimenting with alternative

25

ways to combine the search strategies in HD-DDS is unnecessary. Instead of experimenting with

26

HD-DDS component configurations to improve upon HD-DDS performance for a new WDS

27

problem, it is recommended that HD-DDS users utilize the available time they have for 27

1

performing multiple independent HD-DDS optimization trials or implement alternative local

2

search strategies.

3

In practice, when users apply the HD-DDS algorithm as we suggest, results have shown they

4

will have a HD-DDS solution much more quickly (by 50-70%) than the worst case we

5

recommend they plan for (e.g. under the worst case assumption that all solutions will have their

6

hydraulics evaluated). In deciding how to utilize any available remaining computational budget

7

after their first HD-DDS trial terminates, we suggest that if the user is satisfied with the current

8

HD-DDS solution quality (cost), and HD-DDS terminated before the L2 local searches both

9

converge, we suggest they give HD-DDS the extra time to polish the available solutions with L2.

10

In any other cases (the user is dissatisfied with best cost returned or HD-DDS terminates because

11

all five algorithm steps were completed), we would suggest the user perform a second HD-DDS

12

optimization trial with the remaining budget.

13 14

5.

Conclusions and Future Work

15 16

For the range of WDS benchmark case studies considered in this study, numerical results

17

demonstrate that the HD-DDS algorithm exhibits superior overall performance in comparisons to

18

the MMAS ACO [Zecchin et al., 2007], GENOME GA [Reca and Martinez, 2006], and PSO

19

variant [Montalvo et al., 2008] algorithms.

20

evaluations HD-DDS stochastically dominates MMAS ACO results in all three case studies for

21

which their performance is compared. This is achieved despite the fact that no parameter tuning

22

was conducted in HD-DDS while MMAS ACO parameters were specifically tuned to each case

23

study (involving millions of EPANET2 simulations). The worst HD-DDS result was better than

24

the best GENOME GA result for the 454 decision variable Balerma network even though the

25

GENOME GA utilized 100 times more objective function evaluations. In addition, HD-DDS

26

found a new best solution to the Balerma problem. HD-DDS found the best known solutions

27

more frequently and easily avoided the worst solutions returned by the PSO variant [Montalvo et

For the same number of objective function

28

1

al., 2008] despite the fact that PSO algorithm parameters were determined with preliminary

2

tuning experiments. Furthermore, because the evaluation of many of the candidate solutions

3

identified by HD-DDS does not require simulating network hydraulics (e.g. 50%-70% of

4

objective function evaluations in this study), the actual HD-DDS computation time would be

5

much less (by nearly 50%-70%) than that of the comparative algorithms in this study. This

6

computational advantage of HD-DDS extends over any optimization algorithm requiring that

7

network hydraulics be simulated for all candidate solutions. The parameter-free constraint

8

handling approach based on Deb [2000] was successful.

9

HD-DDS can also be applied to other types of constrained discrete optimization problems in

10

water resources and environmental management such as watershed best management practice

11

optimization [e.g. Arabi et al., 2006], groundwater management and monitoring problems [e.g.

12

Reed et al., 2000], and sorptive barrier design [Matott et al., 2006]. HD-DDS application to new

13

problem types like these requires users consider whether changes are necessary in the two local

14

search types defined here. In the most general interpretation of L1 and L2, HD-DDS is a general

15

methodology that is not specific to WDS problems. L1 enumerates all solutions that differ from

16

the current best solution by a single decision variable (for WDS design this can be very efficient).

17

L2 enumerates all solutions that differ from the current best solution by only two decision

18

variables (again, for WDS design this can be very efficient). Furthermore, for design problems

19

with more complex constraint sets than those considered in this paper, Deb [2000] shows how

20

normalizing all constraint violations enables the use of his constraint handling technique.

21

It is important to note that the benchmark WDS case studies solved here (which minimize

22

cost subject to some design constraints) are gross simplifications of real-world WDS design

23

problems. Walski [2001] discusses why real world problems need to be solved by minimizing net

24

benefits and thus considering multiple different objectives. Our current work is extending the

25

HD-DDS methodology to multi-objective optimization benchmarks that are better representations

26

of real-world WDS design problems. Now that HD-DDS has been shown to be effective relative

27

to benchmark single-objective optimization algorithms, future studies should focus on the 29

1

application of HD-DDS to real-world WDS design problems as formulated by practicing

2

engineers designing the system. In such a study, we would expect that the one-pipe and/or

3

especially the two-pipe change local searches could be modified or replaced with alternate and

4

perhaps case study specific local search strategies that replicate the logic practicing engineers

5

employ when evaluating alternative system designs by trial and error. For example, the two-pipe

6

change local search could be replaced with the grouping method search strategy that Gessler

7

[1985] suggested for large real-world WDS design problems. A more promising and incredibly

8

efficient deterministic local search strategy that could replace or even precede L2 in HD-DDS for

9

polishing discrete DDS solutions, especially for case studies with hundreds of decision variables,

10

is the cellular automaton network design algorithm (CANDA) for WDS optimization introduced

11

by Keedwell and Khu [2006]. CANDA enables expert designers to encapsulate their case study

12

knowledge in the optimization procedure.

13

The HD-DDS algorithm does not require practicing WDS engineers to experiment and

14

identify good optimization algorithm parameters and instead gives them a robust optimization

15

tool with which they can experiment and solve multiple optimization problems that have different

16

design problem characteristics that practicing WDS engineers are ultimately interested in. These

17

would include different design constraint sets, different objectives and different future scenarios

18

leading to different nodal demand scenarios. An additional benefit of HD-DDS is that it scales

19

the search strategy to the user input computational budget.

20

Matlab source codes for HD-DDS are available by emailing the first author and will

21

eventually be available at http://www.civil.uwaterloo.ca/btolson/softare. The EPANET2 input

22

files of the lowest cost solutions found by HD-DDS for each of the five case studies in this paper

23

are available by emailing the first author. Future researchers can replicate our case studies

24

exactly by using these input files in conjunction with the EPANET2 Programmer's Toolkit.

25

30

1

Acknowledgements

2

This research was supported with funding from Dr. Tolson’s NSERC Discovery Grant. We

3

thank the reviewers of this manuscript for their insightful comments which definitely improved

4

the presentation of our findings.

5 6

References

7

Alperovits, E., and U. Shamir (1977), Design of Optimal Water Distribution Systems, Water

8

Resour. Res., 13(6), 885-900.

9

Arabi, M., R. S. Govindaraju, and M. M. Hantush (2006), Cost-effective allocation of watershed

10

management practices using a genetic algorithm, Water Resour. Res., 42, W10429,

11

doi:10.1029/2006WR004931.

12

Bhave, P. R., and V. V. Sonak (1992), A Critical-Study of the Linear-Programming Gradient-

13

Method for Optimal-Design of Water-Supply Networks, Water Resour. Res., 28(6), 1577–

14

1584

15

Broad D.R., Dandy G.C., Maier H.R., and Nixon J.B. (2006) Improving metamodel-based

16

optimisation of water distribution systems with local search. IEEE World Congress of

17

Computational Intelligence, Vancouver, Canada, July 16-21, Proceedings on DVD.

18 19 20 21 22 23 24 25

Cunha, M. d. C., and L. Ribeiro (2004), Tabu search algorithms for water network optimization, Eur. J. Oper. Res., 157(3), 746-58. Cunha, M. d. C., and J. Sousa (1999), Water distribution network design optimization: simulated annealing approach, J.Water Resour. Plann. Manage., 125(4), 215-221. Deb, K. (2000), An efficient constraint handling method for genetic algorithms, Comput. Methods Appl. Mech. Eng., 186(2-4), 311-38. Eiger, G., U. Shamir, and A. Bental (1994), Optimal-Design of Water Distribution Networks, Water Resour. Res., 30(9), 2637–2646.

26

Eusuff, M. M., and K. E. Lansey (2003), Optimization of water distribution network design using

27

the shuffled frog leaping algorithm, J. Water Resour. Plann. Manage., 129(3), 210-225. 31

1

Farmani, R., J.A. Wright, D.A. Savic and G.A. Walters (2005), Self-adaptive fitness formulation

2

for evolutionary constrained optimization of water systems, ASCE Journal of Computing in

3

Civil Engineering, Vol. 19, No. 2, pp. 212-216.

4 5 6 7 8 9 10

Fujiwara, O., and D. B. Khang (1990), A two-phase decomposition method for optimal design of looped water distribution networks, vol.26, no.4 Apr., 539-549, doi: 10.1029/89WR02879. Geem, Z. W. (2006), Optimal cost design of water distribution networks using harmony search, Engineering Optimization, 38(3), 259-280. Gessler, J. (1985), Pipe Network Optimization by Enumeration, Computer Applications in Water Resources, Proceedings of the ASCE Specialty Conference, 1985. Gibbs M.S., Dandy G.C. and Maier H.R. (2008) A genetic algorithm calibration method based on

11

convergence

due

to

genetic

12

doi:10.1016/j.ins.2008.03.012.

drift.

Information

Sciences,

178(14),

2857-2869,

13

Gibbs M.S., Maier H.R. and Dandy G.C. (2009), Comparison of genetic algorithm parameter

14

setting methods for chlorine injection optimization, Journal of Water Resources Planning

15

and Management (accepted, May 13, 2009).

16

Keedwell, E. and ST. Khu (2006), Novel Cellular Automata Approach to Optimal Water

17

Distribution Network Design, Journal of Computing in Civil Engineering, Vol. 20, No. 1:

18

49–56.

19 20 21 22 23 24

Kim, J. H., T. G. Kim, J. H. Kim, and Y. N. Yoon (1994), A study on the pipe network system design using non-linear programming, J. Korean Water Resour. Ass., 27(4), 59–67. Lippai, I., J. P. Heaney, and M. Laguna (1999), Robust water system design with commercial intelligent search optimizers, J. Comput. Civ. Eng., 13(3), 135-143. Lobo, F. G., and Goldberg, D. E. (2004). The parameter-less genetic algorithm in practice. Inf. Sci., 167(1-4), 217-232.

25

Maier, H. R., A. R. Simpson, A. C. Zecchin, W. K. Foong, K. Y. Phang, H. Y. Seah, and C. L.

26

Tan (2003), Ant colony optimization for design of water distribution systems, J. Water

27

Resour. Plann. Manage., 129(3), 200-209. 32

1

Matott LS, Bartelt-Hunt SL, Rabideau AJ. Fowler KR. (2006), Application of heuristic

2

optimization techniques and algorithm tuning to multi-layered sorptive barrier design.

3

Environmental Science and Technology, 40(20), 6354 -6360.

4

Minsker, B. S. (2005). "Genetic Algorithms." Hydroinformatics: Data Integrative Approaches in

5

Computation, Analysis and Modeling, P. Kumar, J. Alameda, P. Bajcsy, M. Folk, and M.

6

Markus, eds., CRC Press, Florida, USA, 439-456.

7

Montalvo, I., J. Izquierdo, R. Pérez, and P. L. Iglesias (2008), A diversity-enriched variant of

8

discrete PSO applied to the design of water distribution networks, Engineering

9

Optimization, 40:7,655-668, doi: 10.1080/03052150802010607

10 11 12 13

Perelman, L., and A. Ostfeld (2007), An adaptive heuristic cross-entropy algorithm for optimal design of water distribution systems, Engineering Optimization, 39(4), 413-428. Prasad T.D. and. Park N.S. (2004), Multiobjective genetic algorithms for design of water distribution networks. J. Water Resour. Plan. Manage, 130(71):73–82.

14

Reed, P. M., B. S. Minsker, and A. J. Valocchi. (2000), Cost effective long-term groundwater

15

monitoring design using a genetic algorithm and global mass interpolation.” Water

16

Resources Research, 36(12), 3731-3741.

17

Reca, J., J. Martinez, C. Gil, and R. Banos (2008), Application of several meta-heuristic

18

techniques to the optimization of real looped water distribution networks, Water Resour.

19

Manage., 22(10), 1367-1379, doi: 10.1007/s11269-007-9230-8.

20 21 22 23 24 25 26 27

Reca, J., and J. Martínez (2006), Genetic algorithms for the design of looped irrigation water distribution networks, Water Resour. Res., 42(5). Rogers, A., and Prugel-Bennett, A. (1999). Genetic drift in genetic algorithm selection schemes. IEEE Trans. Evol. Comp., 3(4), 298-303. Rossman, L. A. (2000), EPANET 2 users manual, Rep. EPA/600/R-00/057, U.S. Environ. Prot. Agency, Cincinnati, Ohio. Savic, D. A., and G. A. Walters (1997), Genetic algorithms for least-cost design of water distribution networks, J. Water Resour. Plann. Manage., 123(2), 67-77. 33

1

Schaake, J. C., and D. Lai (1969), Linear programming and dynamic programming applications

2

to water distribution network design, Rep. No. 116, Hydrodynamics Laboratory, MIT,

3

Cambridge, Mass.

4 5 6 7 8 9

Simpson, A. R., G. C. Dandy, and L. J. Murphy (1994), Genetic algorithms compared to other techniques for pipe optimization, J. Water Resour. Plann. Manage., 120(4), 423-443. Srinivasa, K. G., Venugopal, K. R., and Patnaik, L. M. (2007). A self-adaptive migration model genetic algorithm for data mining applications. Inf. Sci., 177(20), 4295-4313. Suribabu, C. R., and T. R. Neelakantan (2006), Design of water distribution networks using particle swarm optimization, Urban Water Journal, 3(2), 111-120.

10

Tolson, B. A., M. Asadzadeh, A. C. Zecchin, and H. R. Maier (2008), A New Algorithm for

11

Water Distribution System Optimization: Discrete Dynamically Dimensioned Search, paper

12

presented at World Environmental & Water Resources Congress, Honolulu, Hawaii, May

13

12-16.

14 15

Tolson, B. A., and C. A. Shoemaker (2007), Dynamically dimensioned search algorithm for computationally efficient watershed model calibration, Water Resour. Res., 43(1), 01413.

16

Tolson, B. A., H. R. Maier, A. R. Simpson, and B. J. Lence (2004), Genetic algorithms for

17

reliability-based optimization of water distribution systems, J. Water Resour. Plann.

18

Manage., 130(1), 63-72.

19 20 21 22 23 24 25 26

Varma, K. V. K., S. Narasimhan, and S. M. Bhallamudi (1997), Optimal design of water distribution systems using an NLP method, J. Environ. Eng., 123, 381-388. Walski, T. M. (2001), The wrong paradigm - Why water distribution optimization doesn't work, J. Water Resour. Plann. Manage., 127(4), 203-205. Wolpert, D. H., and MacReady, W. G. (1997). No free lunch theorem for optimization. IEEE Trans. Evol. Comp., 1, 66-82. Wu, Z. Y., P. F. Boulos, and C. H. Orr (2001), Using genetic algorithms to rehabilitate distribution systems, American Water Works Association Journal, 93(11), 74-85.

34

1 2

Wu, Z.Y., Simpson, A.R. (2002), A self-adaptive boundary search genetic algorithm and its application to water distribution systems, Journal of Hydraulic Research, 40 (2), 191-203.

3

Zecchin, A. C., H. R. Maier, A. R. Simpson, M. Leonard, and J. B. Nixon (2007), Ant colony

4

optimization applied to water distribution system design: Comparative study of five

5

algorithms, J. Water Resour. Plann. Manage., 133(1), 87-92.

6

Zecchin, A. C., A. R. Simpson, H. R. Maier, M. Leonard, A. J. Roberts, and M. J. Berrisford

7

(2006), Application of two ant colony optimisation algorithms to water distribution system

8

optimisation, Math. Comput. Model., 44(5-6), 451-468.

9

Zecchin, A. C., A. R. Simpson, H. R. Maier, and J. B. Nixon (2005), Parametric study for an ant

10

algorithm applied to water distribution system optimization, IEEE Transactions on

11

Evolutionary Computation, 9(2), 175-191.

12 13 14

35

1 2

Table 1. Characteristics of various optimization algorithms recently applied to WDS optimization.

3 Optimization Algorithm

Application Reference

GA (GENOME)

Reca and Martinez [2006] Montalvo et al. [2008] Eusuff and Lansey [2003] Geem [2006] Reca et al. [2008] Suribabu and Neelakantan [2006]

PSO variant SFLANET HS MSATS7 PSO MMAS ACO CE HD-DDS

4 5 6 7 8 9 10 11 12 13

Zecchin et al. [2007] Perelman and Ostfeld [2007] This study

# of Reported Algorithma + penalty functionb parameters

Do algorithm parameter sets vary by case study?d

Do authors report parameter values determined by case study specific experimentation/optimization?d

8

NO

YES (for 1 of 3 case studies)

8

YES

YES

6

YES

YES

5 5

YES NOc

YES YES

5

YES

YES

4

YES

YES

3

YES

NO

1

NO

NO

a) Algorithm parameter counts do not include stopping criteria for algorithms (such as max. number of generations, max. number of objective function evaluations) since these can be specified based on project specific timelines for when a result is needed. b) All studies (this one excluded) report using a standard penalty function. All studies with a penalty function report either 1 or 2 penalty parameters except Perelman and Ostfeld [2007] who do not report on the form of the penalty function used. c) The authors report that parameters were only transferred from the first to the second case study because the extreme computational burden of the second case study precluded parameter experiments. d) Usually only a subset of reported algorithm parameters are varied/optimized by case study.

14 15 16 17 18

36

1 2 3 4

Table 2. Summary of HD-DDS and other algorithm performance for the five WDS case studies investigated in this study. WDS Case Study

NYTP

HP

NYTP2

BPc

5 6 7 8 9 10 11 12 13

Algorithm (see Table 1 for references)

Objective Function Evaluationsa x number of optimization trials

% of trials with best known solution found

–MMAS ACO

50,000 x 20

HD-DDS

Best Cost (monetary units x 106) and % deviation from best known solution (in brackets) Minimum

Median

Maximum

60

38.638 (0.0)

38.638 (0.0)

39.415 (2.0)

50,000 x 50

86

38.638 (0.0)

38.638 (0.0)

38.769 (0.3)

PSO variant

80,000 x 2000

30

38.638 (0.0)

38.83 (0.5)

47.0b (21.6)

–MMAS ACO

100,000 x 20

0

6.134 (0.9)

6.386 (5.0)

6.635 (9.1)

HD-DDS

100,000 x 50

8

6.081 (0.0)

6.252 (2.8)

6.408 (5.4)

PSO variant

80,000 x 2000

5

6.081 (0.0)

6.31 (3.8)

6.55b (7.7)

CE

97,000 x 1

-

6.081 (0.0)

-

-

GENOME GA

150,000 x 10

10

6.081 (0.0)

6.248 (2.7)

6.450 (6.1)

HD-DDS

150,000 x 50

2

6.081 (0.0)

6.260 (2.9)

6.393 (5.1)

–MMAS ACO

300,000 x 20

5

77.275 (0.0)

78.199 (1.2)

79.353 (2.7)

HD-DDS

300,000 x 20

85

77.275 (0.0)

77.275 (0.0)

77.434 (0.2)

GENOME GA

10,000,000 x 10

0

2.302 (18.7)

2.334 (20.3)

2.35 (21.1)

HD-DDS

100,000 x 10

0

2.099 (8.2)

2.165 (11.6)

2.212 (14.0)

MSATSd

45,000 x 1

-

3.298 (69.9)

-

-

HD-DDS

10,000 x 10

0

2.660 (37.0)

2.759 (42.2)

2.897 (49.3)

a) Unlike all other algorithms, the majority of HD-DDS objective function evaluations do not require evaluating the hydraulics with EPANET2. See discussion in section 4. b) Based on 100 (not 2000) optimization trials (Figure 3 for HP, Figure 5 for NYTP in Montalvo et al. [2008]). c) The best known solution to the Balerma network based on 30 million objective function evaluations is 1.9409 million Euro (see section 3.2.1). d) MSATS (mixed simulated annealing tabu search) is the best of 4 metaheuristics on this problem from Reca et al. [2008].

37

1 2

Figure 1. Discrete Dynamically Dimensioned Search (Discrete DDS) algorithm.

3 4 5 6 7

Figure 2. Example discrete DDS probability mass functions for candidate option numbers (xinew) for a single decision variable with 16 possible options (A and B) and 6 possible options (C and D) under various values for xibest. Default Discrete DDS r parameter of 0.2.

8 9 10 11 12 13 14

Figure 3. Evaluating the objective function in HD-DDS. Note that x is the new solution to be evaluated, xbest is the current best solution, cost(x) calculates the cost of the network based on the diameter and length of pipes, F(x) is the objective function, H(x) is the summation of pressure violations at all nodes in the network, hi(x) is the head at node i, himin is the minimum required head at node i and xmax is the solution with all pipes at their maximum diameter.

15 16 17

Figure 4. The Hybrid Discrete Dynamically Dimensioned Search (HD-DDS) algorithm.

18 19

Figure 5. Progress as of the end of each step of the HD-DDS algorithm (see

20 21 22 23 24 25 26

Figure 4) versus average number of function evaluations for various WDS case studies and corresponding total computational budget input to HD-DDS. 50 optimization trials are shown for each case study except for NYTP2 where only 20 optimization trials are shown. The numbers in brackets count the number of trials where the best solution so far is equal to the best known solution.

27 28 29 30 31 32 33

Figure 6. Empirical CDF of best solutions from HD-DDS and MMAS ACO algorithm for the A) NYTP, B) HP and C) NYTP2 for approximately the same number of objective function evaluations (see the figure for Mtotal in brackets). 20 optimization trials are shown for MMAS ACO. 50 optimization trials for NYTP and HP HD-DDS results are shown. 20 optimization trials for NYTP2 are shown.

34 35 36 37 38

Figure 7. HD-DDS performance with different computational budgets (Mtotal) compared to other algorithm performance on Balerma network. HD-DDS results show all 10 optimization trials. Other algorithm results are for a single trial or show the range of results from multiple trials.

39 40 38

1

39

Tolson, B. A., M. Asadzadeh, H. R. Maier, and A. Zecchin (2009), Hybrid discrete dynamically dimensioned search (HD-DDS) algorithm for water distribution system design optimization, Water Resources Research, 45, W12416, doi:10.1029/2008WR007673. STEP 0. Define discrete DDS inputs which are as follows: • maximum number of objective function evaluations, M • neighborhood perturbation size parameter, r (0.2 is default) • vector with number of discrete options for all D decision variables, xmax. Note that xmin = [1, 1, …, 1] • initial solution, x0 = [x1, …, xD], respecting decision variable bounds STEP 1. Set objective function evaluation counter to 1, i = 1, and evaluate objective function F at initial solution, F(x0): • Fbest = F(x0), and xbest = x0 STEP 2. Randomly select J of the D decision variables for inclusion in neighborhood, {N}: • calculate probability each decision variable is included in {N} as a function of i: P(i) = 1–ln(i)/ln(M) • FOR d = 1, … D decision variables, add d to {N} with probability P(i) • IF {N} empty, select one random d for {N} STEP 3. FOR j = 1, …, J decision variables in {N}, perturb xjbest by sampling from a discrete probability distribution. This discrete distribution approximates a normal probability distribution as follows: • Sample a standard normal random variable, N(0,1) • xjnew = xjbest + σjN(0,1), where σj = r(xjmax - xjmin) • IF xjnew < (xjmin – 0.5), reflect perturbation at xjmin – 0.5: ƒ xjnew = (xjmin - 0.5) + ([xjmin - 0.5] - xjnew) = 2xjmin - xjnew - 1 ƒ IF xjnew > (xjmax + 0.5), set xjnew = xjmin • IF xjnew > (xjmax + 0.5), reflect perturbation at xjmax + 0.5: ƒ xjnew = (xjmax + 0.5) - (xjnew - [xjmax + 0.5]) = 2xjmax - xjnew + 1 ƒ IF xjnew < (xjmin - 0.5), set xjnew = xjmax • Round xjnew to the nearest integer representing the discrete option number • IF xjnew = xjbest, sample xjnew from a discrete uniform probability distribution, U(xjmax, xjmin), until xjnew ≠ xjbest STEP 4. Evaluate F(xnew) and update current best solution if necessary: • IF F(xnew) ≤ Fbest, update new best solution: ƒ Fbest = F(xnew) and xbest = xnew STEP 5. Update objective function evaluation counter, i = i+1, and check stopping criterion: • IF (P(i) < 1/D) OR IF (i = M), STOP, save Fbest & xbest • ELSE, set xnew = xbest, and go to STEP 3

1

Probability

0.16 0.14 0.12 0.10 0.08 0.06 0.04 0.02 0.00

A) xibest = 8

1

0.30

3

4

best

B) xi

0.25 Probability

2

5 6 7 8 9 10 11 12 13 14 15 Option # for Decision Variable xi

= 14

0.20 0.15 0.10 0.05 0.00

0.45 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 0.00

2

3

4

5 6 7 8 9 10 11 12 13 14 15 Option # for Decision Variable xi

C) xibest = 4

1 2 3 4 5 6 Option # for Decision Variable xi

Probability

Probability

1

0.45 0.40 0.35 0.30 0.25 0.20 0.15 0.10 0.05 0.00

D) xibest = 5

1 2 3 4 5 6 Option # for Decision Variable xi

2

Evaluate cost(x) IF [cost(x) > cost(xbest) AND H(xbest) = 0] F(x) = cost(x) (no EPANET run) ELSE evaluate H(x) by running EPANET, then

H( x ) =

∑ max[0,

# nodes i =1

himin − hi ( x )

]

IF H(x) = 0, F(x) = cost(x) IF H(x) > 0, F(x) = cost(xmax) + H(x)

3

STEP 0. Define case study and algorithm inputs: • maximum number of objective function evaluations, Mtotal • current available computational budget, M, as M=Mtotal • network inputs (layout, available pipe diameters, pipe costs, etc.) • vectors of decision variables (pipe diameter option numbers) for the best solutions found so far, xAbest and xBbest defined initially as empty sets, xAbest = [ ], xBbest = [ ] STEP 1. Perform global search with discrete DDS: • Run the Discrete DDS with M as function evaluation limit • Return the best solution of this step, xbest and the corresponding objective function value, F(xbest) • Return the computational effort in this step, m, and update the available computational budget M = M-m STEP 2. Perform a fast local search, L1, which changes current best solution by only one pipe at a time: • Run L1 initialized at xbest • Check if update needed for best solution, xbest, and F(xbest) • Return the computational effort in this step, m, and update the available computational budget M = M-m • IF xAbest is empty, ƒ set xAbest = xbest • ELSE IF xBbest is empty ƒ set xBbest = xbest • IF M ≤ 0, ƒ STOP HD-DDS • ELSE IF xBbest is not empty ƒ Go to STEP 4 • ELSE, Go to STEP 3 STEP 3. Perform the second independent global search with discrete DDS followed by L1: • Go to STEP 1. STEP 4. Perform a slower local search, L2, which changes current best solution by only two-pipes at a time: • Run L2 initialized at the best of xAbest and xBbest (this is referred to as L2') • Check if update needed for xAbest, xBbest, xbest, and F(xbest) • Return the computational effort in this step, m, and update the available computational budget M = M- m • IF M ≤ 0 or xBbest = xAbest ƒ STOP HD-DDS • ELSE, Go to STEP 5 STEP 5. Perform another L2 local search: • Run L2 initialized at the worst of xAbest and xBbest (this is referred to as L2'') • Check if update needed for xAbest, xBbest, xbest, and F(xbest) • Return the computational effort in this step, m and update the available computational budget M = M- m STOP HD-DDS • Return xbest as the best of xAbest and xBbest, corresponding F(xbest) and total objective functions evaluated (Mtotal – M) • Report if xbest is local optimum with respect to L1 or L2.

4

6.8 6.7

STEP1 STEP2 STEP3 STEP4 STEP5 Average

C) HP, Mtotal = 100,000

6.6 6.5 6.4 6.3 6.2 6.1 (0) 6.0 70000

(0)

(4)

(4)

80000 90000 100000 Average number of function evaluations

Objective function (Cost $×10 6)

99 97 95 93 91 89 87 85 83 81 79 77 (0) 75 220000

Objective function (Cost Won×106)

Objective function (Cost $×10 6)

Objective function (Cost $×10 6)

50 A) NYTP, Mtotal = 50,000 STEP1 49 STEP2 48 STEP3 47 STEP4 46 STEP5 Average 45 44 43 42 41 40 39 (4) (5) (29 (43) 38 28000 33000 38000 43000 48000 Average number of function evaluations

B) NYTP2, Mtotal = 300,000

(0) (13)

STEP1 STEP2 STEP3 STEP4 STEP5 Average

(17)

240000 260000 280000 300000 Average number of function evaluations

320000

178.0 D) GYP, Mtotal = 10,000 STEP1 177.9 STEP2 177.8 STEP3 STEP4 177.7 STEP5 177.6 Average 177.5 177.4 177.3 177.2 177.1 177.0 (2) (2) (15) (16) 176.9 7000 8000 9000 10000 11000 Average number of function evaluations

5

Probability of equal or better solution

1.0

A) NYTP, Mtotal = 50,000

0.9 0.8 0.7 0.6 0.5 0.4 0.3

HD-DDS

0.2

MMAS

0.1 0.0 38.5

38.6

38.7

38.8

38.9

39.0

39.1

39.2

39.3

39.4

39.5

6

Probability of equal or better solution

Objective function (Cost $×10 )

1.0

B) HP, Mtotal = 100,000

0.9 0.8 0.7 0.6 0.5 0.4

HD-DDS

0.3 0.2

MMAS

0.1 0.0 6.0

6.1

6.2

6.3

6.4

6.5

6.6

6.7

6

Probability of equal or better solution

Objective function (Cost $×10 ) 1.0 0.9

C) NYTP2, Mtotal = 300,000

0.8 0.7 0.6 0.5 0.4 0.3

HD-DDS

0.2

MMAS

0.1 0.0 77.0

77.5

78.0

78.5

79.0

79.5

80.0

6

Objective function (Cost $×10 )

6

3.8

HD-DDS (1,000) 6

Objective Function (Cost €×10 )

3.6

Reca et al. [2007]

HD-DDS (10,000)

3.4

HD-DDS (100,000)

3.2

HD-DDS (1,000,000) HD-DDS (10,000,000)

3.0

Best HD-DDS + two-pipe

2.8

MSATS

2.6

GENOME GA

2.4

Reca and Martinez [2006]

2.2 2.0 1.8 1.E+03

1.E+04

1.E+05

1.E+06

1.E+07

1.E+08

Number of objective function evaluations

7

Š Start L1 at feasible solution xini with maximum of M objective function evaluations. Š Initialize the following: o current internal best solution, xbest = xini o previous internal best solution to empty set, xpbest = [ ] o iteration (objective function evaluation) counter, i = 0 Š WHILE L1 is not converged (xbest  xpbest) : o xpbest = xbest o FOR j = 1 to D decision variables o xtest = xbest o WHILE pipe j can have diameter decreased from xjbest (and thus a reduced cost): ƒ Decrease the diameter of pipe j by 1 discrete option, xjtest = xjbest-1 ƒ Evaluate objective function, F(xtest) ƒ IF xtest is feasible Š Update xbest by xtest ƒ Update the iteration counter, i = i+1 ƒ IF i = M, STOP L1 ƒ BREAK inner loop when xtest is infeasible because smaller pipe j would also be infeasible Š STOP L1: Return xbest, F(xbest), i, and whether L1 is converged to a local solution such that no further improvement is possible by changing one pipe at a time.

Outline of fast local search L1 which can identify a local minimum such that that no further improvement is possible by changing one pipe at a time.

Š Start L2 at feasible solution xini with maximum of M objective function evaluations. Š Initialize the following: o current internal best solution, xbest = xini o previous internal best solution to empty set, xpbest = [ ] o iteration (objective function evaluation) counter, i = 0 Š WHILE L2 is not converged (xbest  xpbest) and computational budget is not exceeded (i < M): o xpbest = xbest o R = max(x1best, x2best, … , xDbest), which is the maximum number of pipe diameter reductions to consider o FOR r = 1 to R x FOR j = 1 to D decision variables o xtest = xpbest o IF (xjpbest - r) > minimum diameter option number of pipe j (and thus a reduced cost) ƒ FOR k = 1 to D o xtest = xpbest o xjtest = xjpbest – r o IF k  j evaluate increased diameters for pipe k starting at largest diameter Š Set xktest = xkmax + 1 Š WHILE xktest can have diameter decreased by one option and xktest > xkpbest + 1: o Decrease the diameter of pipe k by 1, xktest = xktest – 1 o Calculate cost(xtest). No EPANET run. Note: xtest differs from xpbest in the jth and kth pipes only o IF cost(xtest) < cost(xbest) x Evaluate objective function, F(xtest) x IF xtest is feasible o Update xbest by xtest x Update the iteration counter, i = i+1 x IF i = M, STOP L2 x BREAK inner loop when xtest becomes infeasible because smaller diameters for pipe k would also be infeasible Š STOP L2: Return xbest , F(xbest), i, and whether L2 is converged such that no further improvement is possible by changing two-pipes at a time.

Outline of local search L2 for constrained WDS optimization problem that can identify a local minimum such that no further improvement is possible by changing two decision variables (pipes) at a time.