Abstract
Owing to its inherent difficulty, many heuristic solution methods have been proposed for the capacitated minimum spanning tree problem. On the basis of recent developments, it is clear that the best metaheuristic implementations outperform classical heuristics. Unfortunately, they require long computing times and may not be very easy to implement, which explains the popularity of the Esau and Williams heuristic in practice, and the motivation behind its enhancements. Some of these enhancements involve parameters and their accuracy becomes nearly competitive with the best metaheuristics when they are tuned properly, which is usually done using a grid search within given search intervals for the parameters. In this work, we propose a genetic algorithm parameter setting procedure. Computational results show that the new method is even more accurate than an enumerative approach, and much more efficient.
Similar content being viewed by others
References
Aha DW, Kibler D and Albert MK (1991). Instance-based learning algorithms. Mach Learn 6: 37–66.
Ahuja RK, Orlin JB and Sharma D (2003). A composite very large-scale neighborhood structure for the capacitated minimum spanning tree problem. Opns Res Lett 31: 185–194.
Altınel İK and Öncan T (2005). A new enhancement of the Clarke and Wright savings heuristic for the capacitated vehicle routing problem. J Opl Res Soc 56: 954–961.
Amberg A, Domschke W and Voss S (1996). Capacitated minimum spanning trees: Algorithms using intelligent search. Comb Optim: Theory Pract 1: 9–39.
Bard JF, Huang L, Jaillet P and Dror M (1998). A decomposition approach to the inventory routing problem with satellite facilities. Transportat Sci 32: 189–203.
Battarra M (2010). Exact and heuristic algorithms for routing problems. PhD thesis, Università di Bologna.
Battarra M, Golden BL and Vigo D (2008). Tuning a parametric Clarke-Wright heuristic via a genetic algorithm. J Opl Res Soc 59: 1568–1572.
Bazaraa MS, Sherali HD and Shetty CM (2005). Nonlinear Programming: Theory and Algorithms. Wiley—Interscience: Hoboken, NJ.
Chandran B, Golden BL and Wasil E (2003). A computational study of three demon algorithm variants for solving the traveling salesman problem. In: Bhargava HK (eds). Computational Modeling and Problem Solving in the Networked World. Kluwer Academic Publisher: NY, pp 155–175.
Clarke G and Wright JW (1964). Scheduling of vehicles from a central depot to a number of delivery points. Opns Res 12: 568–581.
Cordeau J-F, Gendreau M, Laporte G, Potvin J-Y and Semet F (2002). A guide to vehicle routing heuristics. J Opl Res Soc 53: 512.
Cortes C and Vapnik V (1995). Support-vector networks. Mach Learn 20: 273–297.
Dai HK and Fujino S (2000). On designing constrained local access networks. In: Sudborough IH and Hsu DF (eds). Proceedings of the 2000 International Symposium on Parallel Architectures, Algorithms, and Networks (ISPAN00). IEEE Computer Society Press: Los Alamitos, CA, USA, pp 167–176.
Daskin MS (2002). Randomized saving algorithms. Private Communication.
Esau LR and Williams KC (1966). On teleprocessing system design, part II: A method for approximating the optimal network. IBM Syst J 5: 142–147.
Gavish B (1991). Topological design of telecommunication networks-local access design methods. Ann Opns Res 33: 17–71.
Golden BL, Pepper J and Vossen T (1998). Using genetic algorithms for setting parameter values in heuristic search. In: Dagli C, Akay A, Buczak M, Ersoy O and Fernandez B (eds). Intelligent Engineering System through Artificial Neural Networks. Vol. 8, ASME Press: NY, pp 239–245.
Hooke R and Jeeves TA (1961). Direct search solution of numerical and statistical problems. J Assoc Comput Mach 8: 212–229.
Morgan J and Sonquist J (1963). Problems in the analysis of survey data, and a proposal. J Am Stat Assoc 58: 415–434.
Nelder JA and Mead R (1964). A simplex method for function minimization. Comput J 7: 308–313.
Öncan T and Altınel IK (2009). Parametric enhancements of the Esau-Williams heuristic for the capacitated minimum spanning tree problem. J Opl Res Soc 60: 259–267.
Papadimitriou CH (1978). The complexity of the capacitated tree problem. Networks 8: 217–230.
Park M-W and Kim Y-D (1998). A systematic procedure for setting parameters in simulated annealing algorithms. Comput Opns Res 24: 207–217.
Parson R and Johnson M (1997). A case study in experimental design applied to genetic algorithms with applications to DNA sequence assembly. Am J Math Manage Sci 17: 369–396.
Patterson R and Pirkul H (2000). Heuristic procedure neural networks for the CMST problem. Comput Opns Res 27: 1171–1200.
Patterson R, Pirkul H and Rolland E (1999). A memory adaptive reasoning technique for solving the capacitated minimum spanning tree problem. J Heuristics 5: 159–180.
Pepper J, Golden B and Wasil E (2002). Solving the traveling salesman problem with annealing-based heuristics: a computational study. IEEE Trans Syst Man Cybern A 32: 72–77.
Quinlan JR (1993). C4.5: Programs for Machine Learning. Morgan Kaufmann: San Francisco, CA.
Robertson W, Golden BL, Runger GC and Wasil E (1998). Neural network models for initial public offerings. Neurocomputing 18: 165–182.
Rumelhart DE, Hinton GE and Williams RJ (1986). Learning internal representations by error propagation. Parallel Distrib Process 1: 318–362.
Uchoa E, Fukasawa R, Lysgaard J, Pessoa A, Poggi da Aragão M and An-drade D (2008). Robust branch-cut-and-price for the capacitated minimum spanning tree problem over a large extended formulation. Math Program, Ser. A 112: 443–472.
Van Breedam A (1996). An analysis of the effect of local improvement operators in genetic algorithms and simulated annealing for the vehicle routing problem. Technical report, RUCA Working Paper 96/14, Faculty of Applied Economics, University of Antwerp, Belgium.
Acknowledgements
The authors acknowledge the supports of Ministero dell’Istruzione, dell’Università e della Ricerca, Italy, Galatasaray University Scientific Research Projects Grant no: 07.402.014, and Boğazici University Scientific Research Projects Grant no: 08A304D.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendix A
Analysis set choice
The strategy used to select the analysis set can affect the performance of a GA. We test four analysis set selection strategies for EWG, EWLS, and EWR.
The first analysis set is obtained through the general and deterministic (G/D) strategy. It is general, because it includes instances from all test problem sets. The first two instances with the two smallest capacities are deterministically selected from the sets cm50, cm100, and cm200, which totally makes six instances. We qualify these three sets as more difficult than the problem sets tc40, tc80, te40, and te80, since the EW and its parametric enhancement end up with larger deviations on them.
The remaining three strategies are dedicated in the sense that they determine the analysis instances particularly from one of the three problem groups tc, te, and cm. We believe that this can give the genetic search procedure a higher chance to learn better the set's structural characteristics, but at the expenses of three executions of the GA. Moreover, we use three rules to determine the dedicated instances in the analysis set. The first rule is deterministic. We choose the first two instances from each one of the two smallest capacity groups, as in G/D; this is what we call D/D. The second rule selects the most difficult instances and it is denoted by D/M. The two instances for which EW3 produces the smallest improvements over EW are selected from the test problem groups. Finally, the third rule is based on the random selection of two instances from each group, and it is referred as D/R. Notice that the size of the analysis sets are four for tc and te, and six for the cm instances, for any analysis set strategy. The five parameter vectors forming the set are listed in Table A1, for the four analysis set selection strategies.
The average percentage deviations and CPU times computed running EWBF3 and EWLS are summarized in Table A2. Note that they are obtained according to four distinct analysis set strategy for 0, 1, 2, 3, 4, and 5 search cubes. The first column denotes the sets of test problem sets. The second column includes the four analysis set selection strategies. Notice that the average relative percent improvements made on EW and total CPU time in seconds are reported; they are computed with the new re-implementations.
The results show that a dedicated analysis set can produce slightly better results than the general analysis set. In some cases, the dedicated algorithm produces even better results than EWBF3; these are marked in bold in Table A2. Moreover, the improvements seem to be larger when the instances solved are large (namely, cm200). The most effective analysis set choice seems to be the D/M, but the results obtained with other strategies are not substantially different.
Moreover, results obtained by increasing the number of cubes considered in the LS step are quite interesting. In fact, the first two cubes introduce major improvements in the solution quality, whereas a higher number of cubes result in marginal improvements for a relatively higher computational cost. This behaviour shows that EWG is able to detect promising areas in parameter space fairly well. Whenever we move unnecessarily far away from these regions, by considering a higher number of cubes, the performance of the method does not improve anymore.
Appendix B
Initialization of the randomized prohibition heuristic
The RP heuristic is effective even by considering a very limited number of EW3 iterations, as reported in Section 4. However, we think we should test the performance of the standalone method and we should verify if the solution quality is dependent or not on the saving list quality.
The first experiment consists in initializing RP with a random saving list. Three random parameters are generated in [0, 2] and the corresponding EW3 saving list initializes 195 iterations of RP. The solution quality obtained is definitely not satisfactory, in fact a single run of EW is on the average better than 195 iterations of the randomly initialized RP algorithm.
On the other hand, if the EW3 saving list is initialized with the EW parameters (ie, α=1, β=γ=0), the RP's average improvement is 3.86% with respect to a single EW execution. If the saving list is computed by considering higher quality parameters, such as the three best parameter vectors obtained after the genetic algorithm (ie, the first three parameter sets in ) and each saving list initializes 195/3=65 RP iterations, the average improvement increases to 4.16%. RP proves to be strongly dependent on the quality of the saving list considered. When the saving list is initialized by even better parameters, as the best parameter set found by EWLS, even with 65 RP iterations are performed, the average improvement is 5.11%. If the parameter vectors are the best three found and 65 RP iterations are run for each corresponding saving list, the average improvement increases to 5.50%.
For the sake of completeness, we pushed this latter configuration by considering a larger number of iterations. We executed 1500 and 3000 iterations of the RP algorithm. The average percentage improvement grows to 5.99% for 3000 iterations, but the improvements beyond 1500 iterations are considerably smaller.
Rights and permissions
About this article
Cite this article
Battarra, M., Öncan, T., Altınel, I. et al. An evolutionary approach for tuning parametric Esau and Williams heuristics. J Oper Res Soc 63, 368–378 (2012). https://doi.org/10.1057/jors.2011.36
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1057/jors.2011.36