Skip to main content
Log in

An evolutionary approach for tuning parametric Esau and Williams heuristics

  • General Paper
  • Published:
Journal of the Operational Research Society

Abstract

Owing to its inherent difficulty, many heuristic solution methods have been proposed for the capacitated minimum spanning tree problem. On the basis of recent developments, it is clear that the best metaheuristic implementations outperform classical heuristics. Unfortunately, they require long computing times and may not be very easy to implement, which explains the popularity of the Esau and Williams heuristic in practice, and the motivation behind its enhancements. Some of these enhancements involve parameters and their accuracy becomes nearly competitive with the best metaheuristics when they are tuned properly, which is usually done using a grid search within given search intervals for the parameters. In this work, we propose a genetic algorithm parameter setting procedure. Computational results show that the new method is even more accurate than an enumerative approach, and much more efficient.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Figure 1
Figure 2

Similar content being viewed by others

References

  • Aha DW, Kibler D and Albert MK (1991). Instance-based learning algorithms. Mach Learn 6: 37–66.

    Google Scholar 

  • Ahuja RK, Orlin JB and Sharma D (2003). A composite very large-scale neighborhood structure for the capacitated minimum spanning tree problem. Opns Res Lett 31: 185–194.

    Article  Google Scholar 

  • Altınel İK and Öncan T (2005). A new enhancement of the Clarke and Wright savings heuristic for the capacitated vehicle routing problem. J Opl Res Soc 56: 954–961.

    Article  Google Scholar 

  • Amberg A, Domschke W and Voss S (1996). Capacitated minimum spanning trees: Algorithms using intelligent search. Comb Optim: Theory Pract 1: 9–39.

    Google Scholar 

  • Bard JF, Huang L, Jaillet P and Dror M (1998). A decomposition approach to the inventory routing problem with satellite facilities. Transportat Sci 32: 189–203.

    Article  Google Scholar 

  • Battarra M (2010). Exact and heuristic algorithms for routing problems. PhD thesis, Università di Bologna.

  • Battarra M, Golden BL and Vigo D (2008). Tuning a parametric Clarke-Wright heuristic via a genetic algorithm. J Opl Res Soc 59: 1568–1572.

    Article  Google Scholar 

  • Bazaraa MS, Sherali HD and Shetty CM (2005). Nonlinear Programming: Theory and Algorithms. Wiley—Interscience: Hoboken, NJ.

    Google Scholar 

  • Chandran B, Golden BL and Wasil E (2003). A computational study of three demon algorithm variants for solving the traveling salesman problem. In: Bhargava HK (eds). Computational Modeling and Problem Solving in the Networked World. Kluwer Academic Publisher: NY, pp 155–175.

    Chapter  Google Scholar 

  • Clarke G and Wright JW (1964). Scheduling of vehicles from a central depot to a number of delivery points. Opns Res 12: 568–581.

    Article  Google Scholar 

  • Cordeau J-F, Gendreau M, Laporte G, Potvin J-Y and Semet F (2002). A guide to vehicle routing heuristics. J Opl Res Soc 53: 512.

    Article  Google Scholar 

  • Cortes C and Vapnik V (1995). Support-vector networks. Mach Learn 20: 273–297.

    Google Scholar 

  • Dai HK and Fujino S (2000). On designing constrained local access networks. In: Sudborough IH and Hsu DF (eds). Proceedings of the 2000 International Symposium on Parallel Architectures, Algorithms, and Networks (ISPAN00). IEEE Computer Society Press: Los Alamitos, CA, USA, pp 167–176.

    Chapter  Google Scholar 

  • Daskin MS (2002). Randomized saving algorithms. Private Communication.

  • Esau LR and Williams KC (1966). On teleprocessing system design, part II: A method for approximating the optimal network. IBM Syst J 5: 142–147.

    Article  Google Scholar 

  • Gavish B (1991). Topological design of telecommunication networks-local access design methods. Ann Opns Res 33: 17–71.

    Article  Google Scholar 

  • Golden BL, Pepper J and Vossen T (1998). Using genetic algorithms for setting parameter values in heuristic search. In: Dagli C, Akay A, Buczak M, Ersoy O and Fernandez B (eds). Intelligent Engineering System through Artificial Neural Networks. Vol. 8, ASME Press: NY, pp 239–245.

    Google Scholar 

  • Hooke R and Jeeves TA (1961). Direct search solution of numerical and statistical problems. J Assoc Comput Mach 8: 212–229.

    Article  Google Scholar 

  • Morgan J and Sonquist J (1963). Problems in the analysis of survey data, and a proposal. J Am Stat Assoc 58: 415–434.

    Article  Google Scholar 

  • Nelder JA and Mead R (1964). A simplex method for function minimization. Comput J 7: 308–313.

    Article  Google Scholar 

  • Öncan T and Altınel IK (2009). Parametric enhancements of the Esau-Williams heuristic for the capacitated minimum spanning tree problem. J Opl Res Soc 60: 259–267.

    Article  Google Scholar 

  • Papadimitriou CH (1978). The complexity of the capacitated tree problem. Networks 8: 217–230.

    Article  Google Scholar 

  • Park M-W and Kim Y-D (1998). A systematic procedure for setting parameters in simulated annealing algorithms. Comput Opns Res 24: 207–217.

    Article  Google Scholar 

  • Parson R and Johnson M (1997). A case study in experimental design applied to genetic algorithms with applications to DNA sequence assembly. Am J Math Manage Sci 17: 369–396.

    Google Scholar 

  • Patterson R and Pirkul H (2000). Heuristic procedure neural networks for the CMST problem. Comput Opns Res 27: 1171–1200.

    Article  Google Scholar 

  • Patterson R, Pirkul H and Rolland E (1999). A memory adaptive reasoning technique for solving the capacitated minimum spanning tree problem. J Heuristics 5: 159–180.

    Article  Google Scholar 

  • Pepper J, Golden B and Wasil E (2002). Solving the traveling salesman problem with annealing-based heuristics: a computational study. IEEE Trans Syst Man Cybern A 32: 72–77.

    Article  Google Scholar 

  • Quinlan JR (1993). C4.5: Programs for Machine Learning. Morgan Kaufmann: San Francisco, CA.

    Google Scholar 

  • Robertson W, Golden BL, Runger GC and Wasil E (1998). Neural network models for initial public offerings. Neurocomputing 18: 165–182.

    Article  Google Scholar 

  • Rumelhart DE, Hinton GE and Williams RJ (1986). Learning internal representations by error propagation. Parallel Distrib Process 1: 318–362.

    Google Scholar 

  • Uchoa E, Fukasawa R, Lysgaard J, Pessoa A, Poggi da Aragão M and An-drade D (2008). Robust branch-cut-and-price for the capacitated minimum spanning tree problem over a large extended formulation. Math Program, Ser. A 112: 443–472.

    Article  Google Scholar 

  • Van Breedam A (1996). An analysis of the effect of local improvement operators in genetic algorithms and simulated annealing for the vehicle routing problem. Technical report, RUCA Working Paper 96/14, Faculty of Applied Economics, University of Antwerp, Belgium.

Download references

Acknowledgements

The authors acknowledge the supports of Ministero dell’Istruzione, dell’Università e della Ricerca, Italy, Galatasaray University Scientific Research Projects Grant no: 07.402.014, and Boğazici University Scientific Research Projects Grant no: 08A304D.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to M Battarra.

Appendices

Appendix A

Analysis set choice

The strategy used to select the analysis set can affect the performance of a GA. We test four analysis set selection strategies for EWG, EWLS, and EWR.

The first analysis set is obtained through the general and deterministic (G/D) strategy. It is general, because it includes instances from all test problem sets. The first two instances with the two smallest capacities are deterministically selected from the sets cm50, cm100, and cm200, which totally makes six instances. We qualify these three sets as more difficult than the problem sets tc40, tc80, te40, and te80, since the EW and its parametric enhancement end up with larger deviations on them.

The remaining three strategies are dedicated in the sense that they determine the analysis instances particularly from one of the three problem groups tc, te, and cm. We believe that this can give the genetic search procedure a higher chance to learn better the set's structural characteristics, but at the expenses of three executions of the GA. Moreover, we use three rules to determine the dedicated instances in the analysis set. The first rule is deterministic. We choose the first two instances from each one of the two smallest capacity groups, as in G/D; this is what we call D/D. The second rule selects the most difficult instances and it is denoted by D/M. The two instances for which EW3 produces the smallest improvements over EW are selected from the test problem groups. Finally, the third rule is based on the random selection of two instances from each group, and it is referred as D/R. Notice that the size of the analysis sets are four for tc and te, and six for the cm instances, for any analysis set strategy. The five parameter vectors forming the set are listed in Table A1, for the four analysis set selection strategies.

Table A1 Results of the genetic search: parameters versus analysis sets

The average percentage deviations and CPU times computed running EWBF3 and EWLS are summarized in Table A2. Note that they are obtained according to four distinct analysis set strategy for 0, 1, 2, 3, 4, and 5 search cubes. The first column denotes the sets of test problem sets. The second column includes the four analysis set selection strategies. Notice that the average relative percent improvements made on EW and total CPU time in seconds are reported; they are computed with the new re-implementations.

Table A2 The effect of the analysis set and the number of search cubes on the performance of EWLS

The results show that a dedicated analysis set can produce slightly better results than the general analysis set. In some cases, the dedicated algorithm produces even better results than EWBF3; these are marked in bold in Table A2. Moreover, the improvements seem to be larger when the instances solved are large (namely, cm200). The most effective analysis set choice seems to be the D/M, but the results obtained with other strategies are not substantially different.

Moreover, results obtained by increasing the number of cubes considered in the LS step are quite interesting. In fact, the first two cubes introduce major improvements in the solution quality, whereas a higher number of cubes result in marginal improvements for a relatively higher computational cost. This behaviour shows that EWG is able to detect promising areas in parameter space fairly well. Whenever we move unnecessarily far away from these regions, by considering a higher number of cubes, the performance of the method does not improve anymore.

Appendix B

Initialization of the randomized prohibition heuristic

The RP heuristic is effective even by considering a very limited number of EW3 iterations, as reported in Section 4. However, we think we should test the performance of the standalone method and we should verify if the solution quality is dependent or not on the saving list quality.

The first experiment consists in initializing RP with a random saving list. Three random parameters are generated in [0, 2] and the corresponding EW3 saving list initializes 195 iterations of RP. The solution quality obtained is definitely not satisfactory, in fact a single run of EW is on the average better than 195 iterations of the randomly initialized RP algorithm.

On the other hand, if the EW3 saving list is initialized with the EW parameters (ie, α=1, β=γ=0), the RP's average improvement is 3.86% with respect to a single EW execution. If the saving list is computed by considering higher quality parameters, such as the three best parameter vectors obtained after the genetic algorithm (ie, the first three parameter sets in ) and each saving list initializes 195/3=65 RP iterations, the average improvement increases to 4.16%. RP proves to be strongly dependent on the quality of the saving list considered. When the saving list is initialized by even better parameters, as the best parameter set found by EWLS, even with 65 RP iterations are performed, the average improvement is 5.11%. If the parameter vectors are the best three found and 65 RP iterations are run for each corresponding saving list, the average improvement increases to 5.50%.

For the sake of completeness, we pushed this latter configuration by considering a larger number of iterations. We executed 1500 and 3000 iterations of the RP algorithm. The average percentage improvement grows to 5.99% for 3000 iterations, but the improvements beyond 1500 iterations are considerably smaller.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Battarra, M., Öncan, T., Altınel, I. et al. An evolutionary approach for tuning parametric Esau and Williams heuristics. J Oper Res Soc 63, 368–378 (2012). https://doi.org/10.1057/jors.2011.36

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1057/jors.2011.36

Keywords

Navigation