Introduction

Traveling salesman problem (TSP) is a classical combinatorial optimization problem that can be described by the weight-directed graph \(G = (\upsilon ,\Omega ,d)\) where \(\upsilon = (1,2,3...,n)\) denotes the set of the nodes or cities, \(\Omega = \{ (i,j)|(i,j) \in \upsilon \times \upsilon \}\) represents the set of the edges, and \(d:\Omega \mapsto {\mathbb{N}}\) is the weight function about each edge (i, j). The goal of the problem is to seek the shortest route visiting every city or node exactly once. The well-known methods to solve TSP include genetic algorithm [1], whale optimization algorithm [2], grey wolf optimization algorithm [3], ant colony optimization [4], etc. Due to the advantage of the positive feedback mechanism and distributed computing, ant colony optimization shows excellent performance in solving TSP.

Ant System algorithm (AS), originally proposed by Italian scholar M.Dorigo in the 1990s, is the classical meta heuristic algorithm which imitated ants foraging in nature [5]. The main idea is that ants can communicate with each other effectively by depositing pheromone on the paths between food source and nest. After the AS algorithm proposed, Dorigo [6] also put forward the Ant Colony System algorithm (ACS) to overcome its premature stagnation problem. There are two main improvements called global updating rule and local updating rule in ACS. According to the global updating rule, only the optimal ant can be allowed to deposit pheromone in each iteration while other ants would diminish the pheromone trail on the tour that they visited in terms of the local updating rule. However, this mechanism can accelerate the convergence of the algorithm effectively at the cost of falling into the local extremum easily yet. In addition, Stützle et al. [7] introduced the Max–Min Ant System algorithm (MMAS) to diversify the solution of the ACO. With the upper and lower restriction in the pheromone matrix, MMAS can avoid algorithm premature stagnation to some certain extent but will be hardly converged when the solutions distribute dispersedly.

To improve the performance of the traditional ACO further, many researchers have improved the algorithm in various aspects. Sangeetha et al. [8] improved ACO with a pheromone enhancement mechanism. It strengthened the attraction of the better path so that reduced useless search. Whereas excessively narrowing the search space would decrease the diversity of the algorithm. Ye et al. [9] used the negative feedback pheromone strategy to expand the search space, which achieved exploration of the ant colony in more unknown areas and avoided the premature stagnation of the algorithm. Comparing the best-so-far path with the iteration-best path, Ning et al. [10] dynamically increased the pheromone concentration on the different edges to enhance the diversity of the algorithm. Although these improvements that explored more searching space could diversify the solutions, the convergence speed of the ant colony would be apt to slow down. To accelerate the speed rate of the algorithm, Guan et al. [11] incorporated an automatic updating mechanism into the ant colony algorithm for constraint satisfaction problem (CSP). The main idea of the improvement is to optimize the suboptimal solutions rather than all solutions so that the convergence speed of the algorithm is improved. Liu et al. [12] adaptively updated the pheromone with iteration-best solution, that is called the dynamic weighted pheromone update mechanism, to accelerate the convergence rate of the algorithm. While the algorithm would easily fall into premature stagnation due to the too-fast convergence speed. In addition, the rational assignment of the parameter for ACO is also a baffling dilemma. To address this conundrum, Mahi et al. [13] incorporated the particle swarm optimization algorithm to optimize the parameters of the ACO, which improved the robustness of the algorithm. Olivas et al. [14] introduced the fuzzy control system to assign the appropriate parameters for the ACO algorithm so that a high-quality solution of the algorithm is achieved. Tuani et al. [15] proposed a novel adaptive parameter adjustment mechanism to improve the adaptability of the algorithm. According to these methods [2, 12, 13], the adaptability of the algorithm has been improved to some extent, but the problems that these algorithms solved are relatively simple and it is difficult to deal with more complex problems. Besides, the ant colony algorithm has achieved promising results in other optimization problems, which embraces scheduling problem [16, 17], robot path planning problem [18], geotechnical problems [19], network routing problem [20], and image segmentation [21], etc.

However, although these methods in the single population above have achieved some progress, the improvements of the single population often strengthen one characteristic as well as weaken another. For example, it will explore more search areas at the expend of increasing search time [9, 10] or will diminish the solution accuracy to accelerate the convergence [11, 12]. In order to balance the functions of the convergence speed and diversity of the algorithm further, the multi-colony algorithm gradually attracts many scholars’ attention. Gambardella et al. [22] studied the multi ant colony algorithm for the first time. They used two ACS colonies to solve vehicle scheduling problems with time windows. Chu et al. [23] tested the performance of seven interaction strategies to control the communication of the homogenous colonies. Twomey et al. [24] analyzed the homogenous multi-ant colony with different communication policies and proposed migrant integration strategy for the interaction. The cooperation on homogenous populations usually only amplifies the single feature in terms of their same characteristics whereas the heterogeneous populations can take full advantage of each other. Dong et al. [25] combined the ant colony algorithm with the genetic algorithm in a novel way to solve the TSPs successfully. Zhang et al. [26] applied two heterogeneous ant colonies to diversify the solution of the algorithm by exchanging the pheromone information. To solve the Vehicle Routing Problem well, Wang et al. [27] incorporated local search into the multi ant algorithm and exchanged the global optimal solution of each colony, which improved the solution accuracy.

On the basis of these above references [22,23,24,25,26,27], the multi-ACO has more competitive than the single colony algorithm in balancing the convergence rate and diversity of the population, while the efficiency of this algorithm also needs to be improved due to the insufficient interaction mechanism. To overcome these shortcomings, some cross-discipline methods are applied to improve the performance of the multi colony algorithm. Yang et al. [28] incorporated the game theory to solve the decision conflicts of each sub-colony, which can promote the efficiency of the multi-colony algorithm. Li et al. [29] used the recommendation algorithm to adapt the communication among populations more accurately.

Compared with a single colony algorithm, the multi-ACO algorithm can achieve better results in small or middle scale TSP instances due to the coordination among multiple colonies. Whereas the communication mechanism, the key of the multi-colony algorithm, is relatively simple, causing that the satisfied results can be hardly achieved when solving large-scale TSP instances. To address this issue, the multi-colony ant optimization with dynamic collaborative mechanism and cooperative game is proposed to strengthen the interactive mechanism and take full advantage of the heterogenous populations. The main contributions and innovations of this research are as follows.

First, to improve the solution accuracy of the algorithm in solving large-scale TSP instances, the cooperative game based on the Shepley value is introduced. It redistributes the pheromone payoff to each participant with respect of their contribution after each iteration so that more useful information can be explored during the searching process and a diversification of the solution can be achieved.

Second, the dynamic collaborative mechanism is applied to improve the interactive mechanism of the heterogenous populations and the mechanism contains two methods. One is called the public path recommendation strategy for MMAS population, which recommends the public paths of ACS populations to MMAS population and improves the astringency of the population. The other is the pheromone fusion mechanism based on information entropy. The aim of this method is to help ACS population jump out of the local extremum by regulating the pheromone distribution when the algorithm falls into stagnation.

Finally, the remainder of this paper are as follows. Section “Related work” reports briefly the basic ACS, MMAS algorithm, information entropy and Shapley value. Section “Proposed algorithm” describes the DCM-ACO algorithm, including dynamic collaborative mechanism and cooperative game model. Section “Experiment analysis” gives the analysis about the performance of the proposed strategies and compares DCM-ACO with the traditional ant colony algorithm and other intelligent algorithms. Section “Conclusion” summarizes and prospects this research.

Related work

Ant Colony system

Ant Colony system (ACS) algorithm mainly contains two-state transition rules. One rule, that the ants construct their paths, is called roulette mode if the random number q is higher than the constant parameter q0, where q and q0 are between [0,1]. The state transition formula is as follows.

$$ P^{k}_{ij} = \left\{ {\begin{array}{*{20}c} {\frac{{\tau^{\alpha }_{ij} \cdot \eta_{ij}^{\beta } }}{{\mathop \sum \nolimits_{{l \in {\text{allowed}}}} \tau^{\alpha }_{il} \cdot \eta_{il}^{\beta } }}{\text{ if }}j \in {\text{allowed}}^{k} } \\ {0{\text{ else}}} \\ \end{array} } \right. $$
(1)

where the \(\tau_{ij}\) represents the pheromone value from city i to city j (vertex (i, j)); \(\eta_{ij} = 1/d_{ij}\) denotes the heuristic information on the vertex (i, j) which dij is the cost of the vertex (i, j); \({\text{allowed}}\) stores a set of cities that the ant k is not visit; \(\alpha\) and \(\beta\) are two weight parameters determining the influence of pheromone value and heuristic information. In addition, \({ }S = {\text{argmax}}\left( {\tau_{ij} \cdot \eta_{ij}^{\beta } } \right)\) is another transition formula that the ants positioned city i move to city j when q < q0.

After finishing a transition from city i to city j, each ant applies a local pheromone update rule to decrease the attraction of the edge (i, j). The formula is as follows:

$$ \tau_{ij} \leftarrow \left( {1 - \xi } \right) \cdot \tau_{ij} + \xi \cdot \tau_{0} $$
(2)

where 0 \(< \xi < 1\) is a pheromone evaporation rate; \(\tau_{0} = 1/\left( {n \cdot l_{n} } \right)\) is the initial pheromone level, where n is the city number and \({\varvec{l}}_{{\varvec{n}}}\) is the tour length created by the nearest neighbour heuristic algorithm.

When all ants complete path construction, the global pheromone updating rule is triggered to add pheromone on the global best tour. The formula is written as:

$$ \tau_{ij} \leftarrow \left( {1 - \rho } \right) \cdot \tau_{ij} + \rho \cdot \Delta \tau_{ij}^{{{\text{bs}}}} $$
(3)
$$ \Delta \tau^{{{\text{bs}}}}_{ij} = 1/L_{{{\text{gb}}}} $$
(4)

where \(0 < \rho < 1\) is pheromone evaporation rate; \(\Delta \tau^{{{\text{bs}}}}_{ij}\) is the value of increasing pheromone and \(L_{{{\text{gb}}}}\) is the length of the best tour.

Max-Min Ant system

To improve the diversity of the traditional ant colony algorithm, Stützle proposed the Max–Min Ant System (MMAS) [7]. The pheromone of MMAS is updated by an alternating iteration-best tour with best-so-far tour in the early run time, which the updating rule is defined as same as formula (3). Besides, the pheromone matrix of the ant colony is limited to the specified range [\(\tau_{{{\text{min}}}}\),\(\tau_{{{\text{max}}}}\)]. If \(\tau_{ij} < \tau_{{{\text{min}}}}\), then \(\tau_{ij} = \tau_{{{\text{min}}}}\); if \(\tau_{ij} > \tau_{{{\text{max}}}}\), then \(\tau_{ij} = \tau_{{{\text{max}}}}\). And it also reinitializes the pheromone matrix to avoid the stagnation. The maximum and minimum values of pheromones are set as follows:

$$ \tau_{\max } = (1/\rho ) \cdot \left( {1/L_{{{\text{gb}}}} } \right) $$
(5)
$$ \tau_{\min } = \tau_{\max } /(2n) $$
(6)

where n is the city number; \(L_{{{\text{gb}}}}\) is the length of the global-best tour.

Information entropy

Information Entropy was the firstly proposed by the American scholar Shannon, which greatly promoted the progress of the information theory [30]. For now, information entropy has been applied in many fields and has achieved good results [31,32,33]. And it shows the effectiveness of information entropy as a measure of the discrete system. the formula is as follows:

$$ E\left( P \right) = - \mathop \sum \limits_{x \in X} P\left( x \right)\log p\left( x \right) $$
(7)

where X is the solution of the problem, P (x) is the probability of x, and \(\mathop \sum \nolimits_{x \in X} P\left( x \right) = 1\).

Shapley value

Shapley value, a way of profit distribution, was originally proposed by scholar Lloyd Shapley in 1953 [34]. The main idea of this method is that the income of each member in the union is proportional to their own contribution. And it has promoted the further development of the cooperative game theory, which the formula is given as follows.

$$ \varphi_{i} (v) = \sum\limits_{s \subset K} {w(s)[v(s) - v(s - i)]} $$
(8)
$$ w(s) = \frac{(s - 1)!(k - s)!}{{k!}} $$
(9)

where the K denotes the union, v represents the payoff function of each sub-union, w(s) is the weight function of each union and \(\user2{\varphi }_{{\varvec{i}}} \left( {\varvec{v}} \right)\) is the payoff function of the i-th player in the union.

Proposed algorithm

In this research, we focus on improving the interactive mechanism and solving the large-scale TSP instances well. The proposed algorithm including cooperative game and dynamic collaborative mechanism is designed. First, in the cooperative game model, a novel pheromone accumulation is designed among ACS sub-colonies. By applying Shapley value, the ant colonies updated the pheromone with respect to their contribution, which more useful knowledge about the problem can be explored during the searching process. Secondly, the dynamic collaborative mechanism, including two methods, is used to improve the interactive mechanism of heterogeneous colonies. By using the public path recommendation strategy, the solution that belongs to the ACS colonies would be strengthened in the MMAS, which can accelerate the convergence speed of the MMAS. While the pheromone matrix of the ACS can be reset when the searching falls into stagnation due to the pheromone fusion mechanism. These two methods can take full advantage of heterogeneous colonies. And the interactive model is shown in Fig. 1.

Fig. 1
figure 1

Dynamic interactive game model

Besides, this part is organized as follows. Section “The cooperative game strategy” is dedicated to applying the Shapley value to the multi-ACO algorithm based on the cooperative game theory. Section“ Dynamic collaborative mechanism” provides the dynamic collaborative mechanism in detail that includes pheromone fusion mechanism and the public path recommendation strategy. Section “Algorithm description” is the algorithm description.

Cooperative game strategy

The cooperative game strategy among ACS populations is mainly to ensure the accuracy of the algorithm. In contrast to other intelligent algorithms, the advantage of the ant colony algorithm is mainly depending on the positive feedback mechanism which performs the solution optimization through accumulating useful information after each iteration. Corresponding to this mechanism in ACO, the global pheromone update rule, as shown in Eqs. (3), (4), plays a key role in ACS. However, too fast convergence speed that created by this mechanism would easily bring premature stagnation of the algorithm. Therefore, a reasonable regulation of pheromone accumulation is the key to the ant colony algorithm.

In this research, the Shapley value based on the cooperative game is introduced to regulate the pheromone accumulation after each iteration. Specifically, before performing the global update rule, we collect the pheromones of each subpopulation so that obtain the total pheromone income b, as shown in the following formula (10). And then the Shapley formula is introduced to allocate those total income to each member with respect of their contribution, which is shown in Fig. 2. The aim of this improvement is to regulate the pheromone accumulation of each subpopulation so that makes the pheromone allocation more reasonable, which achieves a diversification of research. The specific distribution formulas are constructed as follows.

$$ b = \sum\limits_{i = 1}^{k} {\Delta \tau_{{{\text{ACS}}}}^{i} } $$
(10)
$$ {\text{Contribution}}_{i} = {\text{sol}}_{i} \cdot H(p_{i} ) $$
(11)

and \({\text{sol}}_{i} { = }\frac{{{\text{length}}_{{{\text{min}}}} }}{{{\text{length}}_{i} }}\), \(H(p_{i} ) = \frac{{E(p_{i} )}}{{E(p_{\max } )}}\). where lengthmin and \(E({p}_{\mathrm{max}})\) are the minimum length and the maximum value of information entropy respectively among k colonies of ACS, and \(so{l}_{i}\) denotes the tour length ratio, \(H({p}_{i})\) represents the diversity ratio of each subpopulation. It can be concluded that the contribution of the colony is the value between [0,1], which the closer the value is to 1, the larger the contribution of the population. And the pheromones are redistributed as follows.

$$ \pi_{i} = \frac{{{\text{Contribution}}_{i} }}{{\sum\nolimits_{i = 1}^{k} {{\text{Contribution}}_{i} } }} $$
(12)
$$ \Delta \tau_{i}^{{{\text{new}}}} = \pi_{i} \cdot b $$
(13)

where \(\Delta {\tau }_{i}^{\mathrm{new}}\) is the new pheromone revenue of the i-th ACS population.

Fig. 2
figure 2

Cooperative game strategy

From equal (11), it can be concluded that in the initial stage of the algorithm, each population is in an exploration state causing that the difference of diversity among sub-colonies is not obvious, thus the contribution of the population is proportional to the quality of its solution. While in the late stage that is called the exploitation stage, the difference of the solution quality has not competitiveness at this stage, so that the contribution of the population is mainly depending on the diversity, corresponding to the information entropy in this research. Besides, the role of the pheromone redistribution strategy, as shown in equal (12), is to regulate the pheromone accumulation of each sub-colony with respect to their contribution in different stages. The aim is to reward the elite population and punish the poor performance population and thus to diversify the solution of the algorithm effectively.

Dynamic collaborative mechanism

In this part, we apply a dynamic collaborative mechanism to take full advantages of the heterogenous population. And the mechanism contains two methods: one is the pheromone fusion mechanism which fuses the pheromone of ACS population and MMAS population based on information entropy to help the algorithm get rid of local minima, and the other is public path recommendation strategy that recommends the public paths from ACS population to MMAS population to improve the convergence rate of the MMAS population.

Pheromone fusion mechanism

Diversity description

On the basic of part 2.3, the information entropy is worthy of measuring the diversity of the algorithm. And the diversity of the population based on information entropy can be described as follows:

$$ P_{i} \left( t \right) = n/m $$
(14)
$$ E\left( {{\text{Pt}}} \right) = - \mathop \sum \limits_{i = 1}^{m} P_{i} \left( t \right)\log p_{i} \left( t \right) $$
(15)

where \(P_{i} \left( t \right)\) is the proportion that the i-th tour trail selected by n ants when M ants generate m paths in this iteration, \(E\left( {{\text{Pt}}} \right)\) is the information entropy of the population in t-th iteration which demonstrates that if the tour difference of the population is higher, the information entropy of the population will be larger, and vice versa. In another word, the higher information entropy, the better diversity of the population. Therefore, by detecting the information entropy of each subpopulation, the state of the algorithm can be described more accurately, thus the adaptability of interaction is improved.

Pheromone fusion strategy

This strategy proposed in this part is main for ACS colonies. Due to the pheromone update mechanism of MMAS, the pheromone distribution in the population has been limited in a certain range, which prevents the population from falling into local optimum effectively. While ACS will easily fall into premature stagnation in terms of the global updating rule although it can speed up the convergence of the algorithm. Under these circumstances, the pheromone fusion strategy is proposed to regulate the pheromone distribution in ACS and diversify the solution of the population, which helps ACS get rid of the local extremum. When the entropy of the ACS colony is lower than the threshold (\(E({p}_{i})<{E}^{*}(P)\), where the \({E}^{*}(P)\) is the threshold parameter), that means the diversity of the population has been decreased, then the pheromone fusion strategy is triggered. And the formula is as follows.

$$ {\text{ph}}_{{{\text{ACS}}}}^{i} \leftarrow (1 - \omega_{i} ) \cdot {\text{ph}}_{{{\text{ACS}}}}^{i} + \omega_{i} \cdot {\text{ph}}_{{{\text{MMAS}}}} $$
(16)

where Phi is the pheromone matrix of population i, Phj is the pheromone matrix of population j; Wj is the pheromone contribution of population j to population i, which the formula can be written as follows.

$$ \omega_{i} = \frac{{E(p)_{{{\text{ACS}}}}^{i} }}{{E(P)_{{{\text{ACS}}}}^{i} + E(P)_{{{\text{MMAS}}}} }} $$
(17)

where E(Pi) is the information entropy of the population i.

Public path recommendation strategy

The strategy in this section is mainly for MMAS population. Although the MMAS population usually has good diversity due to the pheromone restriction, the slow convergence speed is a significant disadvantage of the colony. To address this issue, we propose a public path recommendation strategy, as shown in Fig. 3, to recommend the excellent paths from ACS population to MMAS population, which improves the astringency of the population. Generally speaking, it can be considered that the path selected by multiple populations together is a part of the optimal global path or there is an optimal path near this path. Therefore, we recommend the common paths among ACS populations to MMAS population and reward the pheromone on them when the convergence speed of MMAS population is lower. Thus, it will avoid the useless search of population, save the cost of the solution and accelerate the convergence speed of the algorithm. And the formulas are constructed as follows.

$$ {\text{con}}_{t} { = }\frac{{{\text{iter}}_{{{\text{opt}}}} }}{{{\text{iter}}_{t} }} $$
(18)

where cont represents the convergence rate of the population at the t-th iteration, \({\mathrm{iter}}_{\mathrm{opt}}\) is the number of iterations corresponding to the current optimal solution of the population, and \({\mathrm{iter}}_{t}\) denotes the number of the t-th iteration. From the formula, its range is [0,1] and the closer its value is to 1, the faster convergence speed of the population. And the public path recommendation strategy is triggered when the convergence rate of the MMAS is lower than threshold (cont < con*), which can be written as follows.

$$ \tau_{{{\text{public}}}} = \frac{1}{n} \cdot e^{{ - {\text{iter}}}} $$
(19)
$$ \tau_{ij}^{{{\text{public}}}} \leftarrow \tau_{ij}^{{{\text{public}}}} + \tau_{{{\text{public}}}} $$
(20)

where \( \tau_{{{\text{public}}}}\) is the added pheromone of the common route between ACS populations, n is the city number, iter is the number of the iteration. and in formular (14), \(\tau_{ij}^{{{\text{public}}}}\) denotes the pheromone on the paths that belongs to the common part of ACS populations.

From the formula (13) and (14), we can see that the pheromones are rewarded decreasing with the run of iteration, which accelerates the convergence speed of MMAS population in the early state and improve the diversity in the late stage of the algorithm.

Fig. 3
figure 3

Public path

Algorithm description

The proposed algorithm in this research mainly contains two ACS populations and one MMAS population and each subpopulation construct the path in parallel. Firstly, initializing the parameters among ACS and MMAS colonies in the initial stage of the algorithm. Then, the dynamic collaborative mechanism is triggered to judge the state of each colony. If the information entropy of ACS population is lower than the threshold, the pheromone fusion mechanism is introduced to regulate their pheromone distribution and help the algorithm jump out of the local optimum. Otherwise, the cooperative game strategy is applied to re-update the global pheromone among the ACS populations by the Sharpley value, which enhances the diversity of the algorithm. Meanwhile, the public path recommendation strategy is utilized when the convergence rate of MMAS population is too low. And it recommends the public paths among ACS populations to MMAS population to accelerate its convergence. And the algorithm flowchart is shown in Fig. 4 in detail.

Fig.4
figure 4

The flow chart of DCM-ACO algorithm

figure a

Experiment analysis

In this part, we implement the simulation experiment to verify the performance of the proposed algorithm. The experimental platform is the MATLAB R2019b in Windows 10 environment, the CPU, with 16 GB RAM memory capacity, is Intel(R) Core (TM) i7-10700F, and the experiments are applied to execute based on different scale TSP instances, each instance runs for 20 times independently. In addition, the rest of this part is as follows. Section “Parameters setting” selects the appropriate parameters for ACS, MMAS and DCM-ACO. Section “Strategy analysis” is given the analysis about the strategy we have proposed. Section “Comparison with traditional ACO algorithm” compares the proposed algorithm with the conventional ACO algorithms. Section 4.4 is the experimental comparison between DCM-ACO and other swarm intelligence algorithms. Section “Statistical test analysis” is the statistical test analysis of the experimental results.

Parameters setting

The first part of this section is the traditional parameters setting. And we apply the orthogonal tests that contain four levels and five factors to set the appropriate value for those parameters. Besides, the kroB100 instance is adopted to carry out for each colony in the orthogonal test. Each combination of the parameters experiment is tested 20 times independently to ensure the reliability of the experiment. Tables 1, 2, 3 are the experimental results of ACS and Tables 4, 5, 6 denote the results of MMAS.

Table 1 Experimental factors and levels of ACS
Table 2 Orthogonal test scheme and test results of ACS
Table 3 Analysis of test results of ACS
Table 4 Experimental factors and levels of MMAS
Table 5 Orthogonal test scheme and test results of MMAS
Table 6 Analysis of test results of MMAS

According to the sum of experimental results at each level Hi and the average of each level hi, the parameters α, β, ρ and q0 obtain more excellent results at Level 1, Level 3, Level 1, Level 3, and Level 3 respectively in the orthogonal test of ACS, which are shown the bold numbers in Table 3. While in the orthogonal test of MMAS, the experimental results of the factors α, β and ρ at Level 1, Level 4 and Level 1 respectively, corresponding to the bold numbers in Table 6, are superior to them at other Levels. Therefore, the optimal parameter of ACS algorithm is:\(\alpha = 1,\beta = 4,{ }\rho = 0.1,\xi = 0.3,q0 = 0.8, \) the optimal parameter of MMAS algorithm is: \( \alpha = 1,\beta = 5,\rho = 0.1\)

The second part of this parameter experiment is the newly proposed parameter setting which concludes the information entropy threshold (E(p)*) and convergence rate threshold (con*). In this paper, we select the suitable value of the newly proposed parameter through the experiment of the different values. Besides, the kroB100 and kroA200 instances are selected to carry out the experiment, which the results are shown in Figs. 5 and 6. From the experimental results, the minimum fitness function is obtained under the case that E(p)* = 4 and con* = 0.8.

Fig.5
figure 5

Adjustment of the entropy threshold

Fig.6
figure 6

Adjustment of the convergence threshold

From the above experimental results, the final setting results of the algorithm parameters are shown in the following Table 7, which the ρ denotes the global pheromone evaporation rate, ζ is the local pheromone evaporation rate and M represents the number of ants.

Table 7 The parameter setting of the algorithm

Strategy analysis

In the first phase of the experiment, we analyse the effectiveness of three strategies proposed above including cooperative game mechanism, pheromone fusion mechanism and public path recommendation strategy. LOST-1 is the algorithm that has a pheromone fusion mechanism and public path recommendation strategy but does not use a cooperative game mechanism. LOST-2 is the algorithm that retains cooperative game mechanism and public path recommendation strategy but does not use pheromone fusion mechanism. LOST-3 represents the algorithm that has a cooperative game mechanism and pheromone fusion mechanism but does not use a public path recommendation strategy. In the experiment, kroB100 and lin318 TSP instances are selected and analysed with three aspects including optimal solution error rate, worst solution and average solution. And each instance runs 20 times, 2000 iterations each time. The experimental results are shown in Table 8 and Fig. 7.

Table 8 Experiment results with a different strategy
Fig. 7
figure 7

Comparison results in different strategy

From the experimental results, in kroA100 instance, all the experimental groups gain the optimal solution, and in lin318 instance the optimal solution of the DCM-ACO outperforms other algorithms. Besides, the average solution and worst solution of LOST-3 are superior to the LOST-1 and LOST-2, which proves that the cooperative game strategy and the pheromone fusion mechanism can enhance the stability of the algorithm and improve the solution quality. But from Fig. 7, the convergence speed of LOST-3 is slower than other comparison algorithms showing that the public path recommendation strategy can accelerate the convergence of the population. In a word, the three novel methods we proposed above can improve the performance of the algorithm in various aspects.

Comparison with traditional ACO algorithm

In this part, we select different TSP instances to verify the performance of the proposed algorithm in this research compared with traditional ACOs and the analysis aspects include the best solution, the worst solution, mean solution, error rate and the standard deviation, which the error rate and standard deviation formula are as follows:

$$ {\text{Error}} = \frac{{L_{{{\text{ACO}}}} - L_{{{\text{opt}}}} }}{{L_{{{\text{opt}}}} }} \times 100\% $$
(21)

where LACO represents the optimal solution of each algorithm, and Lopt represents the standard optimal solution of the known test set.

$$ {\text{std}} = \sqrt {\frac{1}{N}\mathop \sum \limits_{i = 1}^{N} \left( {L_{i} - \overline{L}} \right)^{2} } $$
(22)

where std is the standard deviation, N denotes the number of times the algorithm runs, and Li represents the solution obtained by the algorithm in the i-th experiment.

Table 9 shows the experimental results of ACS, MMAS and DCM-ACO algorithm. In small-scale instances of less than 300 cities, although the gap of the error rate between DCM-ACO and the comparison algorithm is not obviously, the standard deviation in DCM-ACO is superior to ACS and MMAS due to the coordination of multiple colonies, showing that our proposed algorithm has a more stable ability. When solving middle-scale instances (city scale within 600), DCM-ACO still remains with higher solution accuracy. And in large-scale TSP instances from 1000 to 2000 cities, the search space is so complex that traditional ant colony algorithms can hardly deal with well, corresponding to the huge error rate of ACS and MMAS in Table 9. However, the error computed based on DCM-ACO is controlled within 2%. These excellent results are mainly ascribed to the cooperative game strategy. With the help of this strategy, the pheromone accumulation of populations after each iteration updates more adaptively with respect to their contributions, so that a diversification of the search is achieved. As a result, the performance of the algorithm is improved especially for large-scale instances. For example, in d2103 instance, ACS and MMAS have huge deviations that the error rate is 4.00% and 5.73% respectively while the error rate found by DCM-ACO is just 1.87%, which greatly outperforms the traditional ACO algorithms. Besides, attribute to the dynamic collaborative mechanism, the interactive frequency of heterogeneous populations is more adaptive and the efficiency of the communication among sub-colonies can be improved greatly so that a strong robustness of the algorithm is acquired. And from the average solution and standard deviation, all the values of these factors obtained by DCM-ACO from small-scale instances to large-scale instances are lower than that based on the comparison algorithms, which proves the strong stability of the proposed algorithm for TSP.

Table 9 Experimental results of DCM-ACO

Figure 8 illustrates the evolution of the best solution calculated with ACS, MMAS and DCM-ACO in the optimization process for pr439, p654 and fl1400. Due to the cooperative game strategy, the higher quality of the solution can be strengthened in the early stage, while the information with better diversity plays an important role in the late stage. It means that the improved algorithm has better exploration capacity in the initial stage and owns the strong exploitation ability in the late stage. As shown in Fig. 8, DCM-ACO outperforms ACS and MMAS in the initial stage. And there is a tremendous gap between the results calculated by these three algorithms. According to the dynamic collaborative mechanism, the advantages of heterogeneous populations can be fully exploited. Due to the public recommendation strategy, the convergence of MMAS can be accelerated and with the help of the pheromone fusion mechanism, ACS can jump out of the local optima effectively. As we can see in Fig. 8, for each instance, DCM-ACO completes the convergence in a short cycle whereas ACS and MMAS converge to the best solution in a larger number of iterations, and the solution accuracy is also superior to the comparison algorithm. In contrast to ACS and MMAS, DCM-ACO can accelerate the convergence without losing high-quality solutions.

Fig. 8
figure 8

Comparison the convergence of different algorithms

In a word, from those results, it can be concluded that DCM-ACO can balance the convergence speed and the diversity of the algorithm. In addition, to verify the authenticity of the experimental results, the optimal tours found by CGMACO are shown in Fig. 9.

Fig. 9
figure 9

Optimal tour found by DCM-ACO

Comparison with other algorithms

The first phase of this section is the comparison of the proposed algorithm with SOS-ACO [4] and PSO-ACO-3opt [13], which SOS-ACO is the ACO with hybrid symbiotic organisms search algorithm and PSO-ACO-3opt is the hybrid method based on Particle Swarm Optimization, Ant Colony Optimization and 3-Opt algorithms. And the bold number in Tables 10 and 11 denotes the minimum value of each index calculated by these algorithms.

Table 10 Compare DCM-ACO with SOS-ACO
Table 11 Compare DCM-ACO with PSO-ACO-3opt

In the experiment of the comparison with SOS-ACO, we analyse the experimental results under three factors that include the best solution, average solution and standard deviation, which the results are shown in Table10. In eil51, eil76 and kroA100 instances, both DCM-ACO and SOS-ACO find the optimal solution whereas DCM-ACO is superior to the SOS-ACO in the rest of instances. Moreover, the average solution computed with DCM-ACO outperforms that computed by SOS-ACO in all instances. Besides, DCM-ACO has a better standard deviation than SOS-ACO in the 5 instances based on 8 instances. In the comparison experiment of PSO-ACO-3opt, three indexes that contain the best solution, worst solution and average solution are selected to analyse the experimental results which are given in Table 11. From Table 11, the average solution calculated by PSO-ACO-3opt is superior to DCM-ACO in eil51 and eil76. However, DCM-ACO has a better average solution than PSO-ACO-3opt in 6 other instances, such as kroA100, ch150, kroA200, pr264, fl417, pr439 and p654. In addition, DCM-ACO outperforms the comparison algorithm in most instances under the best solution and the worst solution, especially in pr439 and p654. In total, the experimental results show that DCM-ACO has strong competitiveness with comparison algorithms for TSP.

In the second phase of this section, we compare the DCM-ACO with various other optimization algorithms. The comparison optimization algorithms are mainly improved ant colony algorithms that include HAACO [15], PACO-3opt [35], HMMA [36], NACO [28], LDTACO [29], PPACO [37] and other swarm intelligence algorithms that include IVNS [38], Discrete Bat Algorithm DBAL [39], Discrete Spider Monkey Optimization DSMO [40], Discrete Symbiotic Organisms Search algorithm DSOS [41] and reference [42]. Tables 12 and 13 show the specific experiment data in small-scale and large-scale TSP instance respectively. And the Best denotes the best solution obtained by each algorithm, the error is the error rate defined by equal (21). And the “-” in the table denotes that the comparison algorithm does not test the instance.

Table 12 Compare proposed algorithm with other algorithms in small-scale TSP instances
Table 13 Compare proposed algorithm with other algorithms in large-scale TSP instances

In small-scale TSP instances, as shown in Table 12, our proposed algorithm finds all the standard optimal solutions from eil51 to pr264, which outperforms the comparison algorithms such as LDTACO, DSMO, PACO-3opt and HAACO. Moreover, in Table 13, which shows the experimental results in large-scale instances, DCM-ACO is also superior to the recent algorithms. For the lin318 and fl417 instances, DBAL find the minimum value over all the six competing algorithms. While in other instances such as pr439, p654, rl1323 and fl1400, DCM-ACO obtains better solutions over all the comparison algorithms. Especially for large-scale TSP instances, the solution computed by DCM-ACO is superior to that found by IVNS and reference [42] in rl1323. Besides, in fl1400, our algorithm also outperforms IVNS, HMMA and reference[42]. In the light of these results, DCM-ACO has strong competitiveness with the state-of-art algorithms for TSP, especially for large-scale instances.

Statistical test analysis

To verify the effectiveness of the proposed algorithm, we perform the Wilcoxon rank-sum test in this section. In this test, P-value denotes the significance level, and if P-value < 0.05, the sig in Tables 14, 15 is YES, which means that the original hypothesis is rejected and the performance between our proposed method and the comparison algorithms is significantly different. Otherwise, the sig is NO, which implies that there is no significant difference among these algorithms. In this part, the original hypothesis is given as: the performance between the algorithms is no significant difference. From Tables 14, 15, all the values of P-value in two tests are less than 0.05, it can be concluded that the comparison experiments between DCM-ACO and traditional ant colony algorithms have statistical significance.

Table 14 Rank sum test between DCM-ACO and ACS
Table 15 Rank sum test between DCM-ACO and MMAS

In addition, to require a fair conclusion further, we perform Friedman’s test to verify the significant difference between our proposed method and other improved algorithms including NACO [28], LDTACO [29] and PPACO [37]. First, we set the null hypothesis H0 is: DCM-ACO has no significant difference with NACO, LDTACO, and PPACO. Then 10 experimental data from the kroA200 and lin318 instances are selected to carry out Friedman’s test under the SPSS25 software. The results are shown in Fig. 10 and Table 16. As we can see in Fig. 10, the decision is that rejection for the H0 which means the proposed method has a significant difference with comparison algorithms. Specifically, fromFig.11, the mean rank of DCM-ACO, NACO, LDTACO and PPACO is 1.27, 3.07, 2.03 and 2.93 respectively. Since the response rates differ at different frequencies, pairwise comparisons are necessary. Table 16 shows the results of the pairwise comparison test. From Table 16, the Adjust significance of DCM-ACO with NACO, LDTACO and PPACO is 0.011, 0.002, 0.001, respectively, which all the values are less than 0.05. It means that DCM-ACO is different from NACO, LDTACO and PPACO. In a word, the comparison experiments between DCM-ACO and other algorithms have statistical significance.

Fig. 10
figure 10

Hypothesis test summary

Fig. 11
figure 11

Related-samples Friedman’s Two-way analysis of variance by ranks

Table 16 Comparison test

Conclusion

In this paper, we have proposed a novel ant colony algorithm, so-called multi-ACO with dynamic collaborative mechanism and cooperative game, to solve traveling salesman problems. In the proposed algorithm, two colonies of Ant Colony System (ACS) and one colony of Max–Min Ant System (MMAS) are formed heterogeneous population. And two ACS populations can better amplify the convergence speed of the algorithm. One MMAS subpopulation is added to enhance the diversity of the population effectively. The advantages of multiple populations complement to ensure the solution quality of the algorithm.

In addition, the cooperative game based on Shapley value is designed to regulate the pheromone accumulation among ACS subpopulations with respect to the contribution of each participant after each iteration. From the experimental results, it portrays that the cooperative game has been proved to be effective and it can adaptively control the pheromone accumulation and improve the diversity of the algorithm.

The dynamic collaborative mechanism is introduced to coordinate among heterogeneous populations. Two methods that pheromone fusion mechanism and public path recommendation strategy are in the mechanism. The former method based on information entropy can help ACS population get rid of the local optimal and the latter accelerates the convergence of the MMAS population effectively. The experiment results in different scale TSPs demonstrate that the improved algorithm can fully complement the advantages of the heterogeneous populations and balance the convergence speed and the diversity of the algorithm effectively, especially for large-scale TSP instances.

In the future, more types of collaborative mechanism can be designed in heterogeneous populations and more kinds of payoff distribution strategies based on game theory can be used to promote co-evolution among populations. In addition, except for the evaluation criteria under information entropy in this paper, other cross-discipline methods such as statistics or machine learning can also be attempted to improve the performance of the population further. Finally, the game mechanism we proposed in this research also has some certain practical value in the application of the ant colony algorithm.