Table of Contents Author Guidelines Submit a Manuscript
Computational Intelligence and Neuroscience
Volume 2016, Article ID 1510256, 12 pages
Research Article

Solving the Traveling Salesman’s Problem Using the African Buffalo Optimization

Faculty of Computer Systems & Software Engineering, Universiti Malaysia Pahang, 26300 Kuantan, Malaysia

Received 16 April 2015; Revised 2 August 2015; Accepted 27 August 2015

Academic Editor: Hong Man

Copyright © 2016 Julius Beneoluchi Odili and Mohd Nizam Mohmad Kahar. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


This paper proposes the African Buffalo Optimization (ABO) which is a new metaheuristic algorithm that is derived from careful observation of the African buffalos, a species of wild cows, in the African forests and savannahs. This animal displays uncommon intelligence, strategic organizational skills, and exceptional navigational ingenuity in its traversal of the African landscape in search for food. The African Buffalo Optimization builds a mathematical model from the behavior of this animal and uses the model to solve 33 benchmark symmetric Traveling Salesman’s Problem and six difficult asymmetric instances from the TSPLIB. This study shows that buffalos are able to ensure excellent exploration and exploitation of the search space through regular communication, cooperation, and good memory of its previous personal exploits as well as tapping from the herd’s collective exploits. The results obtained by using the ABO to solve these TSP cases were benchmarked against the results obtained by using other popular algorithms. The results obtained using the African Buffalo Optimization algorithm are very competitive.

1. Introduction

The growing need for profit maximization and cost minimization has never been greater in human history than what we have today. This need has made optimization a very favoured area of scientific investigations. This development has led to the design of a number of optimization algorithms. Some of the most popular algorithms are the Particle Swarm Optimization [1], Ant Colony Optimization [2], Genetic Algorithm [3], Artificial Bee Colony [4], and many others. However, the above algorithms do have some drawbacks ranging from premature convergence [5], delay in obtaining results, easily being stuck in local minima, and complicated fitness function to having many parameters that require setting up [6]. An attempt to proffer solutions to some of the weaknesses of these algorithms is the motivation for the development of the African Buffalo Optimization (ABO).

The ABO is a population-based stochastic optimization technique that has its inspiration from the behaviour of African buffalos: a species of wild cows, similar to their domestic counterparts that navigate their way through thousands of kilometres in the African rain forests and savannahs by moving together in a large herd of, sometimes, over a thousand buffalos. Their migration is inspired by their search for lush grazing lands. They tend to track the movement of the rainy seasons when they can get lush grazing pastures. As the seasons differ from location to location in the vast African landscape, the buffalos are always mobile in pursuit of their target pastures. In ABO, our interest is in how the buffalos are able to organize themselves in searching the solution space with two basic modes of communications, that is, the alarm “waaa” sounds to indicate the presence of dangers or lack of good grazing fields and, as such, asking the animals to explore other locations that may hold greater promise. On the other hand, the alert “maaa” sound is used to indicate favourable grazing area and is an indication to the animals to stay on to exploit the available resources.

The Traveling Salesman Problem. The Traveling Salesman’s Problem (TSP) is the problem faced by a salesman who, starting from a particular town, has the assignment of finding the shortest possible round trip through a given set of customer towns or cities. The salesman has a mandate to visit each city once before finally returning to the starting town/city. The Travelling Salesman’s Problem (TSP) can be represented by a complete weighted graph with being the set of nodes (cities) and being the set of edges fully connecting the nodes in the graph . In this graph, each edge is given a weight which represents the distance between cities and . It may be important to emphasize that the distance between towns/cities may be symmetric (where the distances between the cities are the same in either going or returning from the towns/cities) or asymmetric (where due to some restrictions, possibly, due to one-way lanes or other reasons, the distances of going from city to city may not be the same). The basic minimization equation for TSP is given towns and their coordinates, find an integer permutation with being the city “”, our task is to minimize the sum of the cities [7, 8]. ConsiderHere, represents the distance between city “” and city and is the distance of city “” and city “.” The aim is to find the shortest path between the adjacent cities in the form of a vector of cities.

This paper is organized in this way: Section 1 introduces the African Buffalo Optimization and the Travelling Salesman’s Problem; Section 2 presents a brief review of relevant literature; Section 3 discusses the basic flow of the African Buffalo Optimization (ABO) algorithm; Section 4 is concerned with how the ABO solves the TSP; Section 5 deals with the experiments and discussion of the results of the symmetric TSP instances; Section 6 is concerned with the experiments on the asymmetric TSP instances and the discussions of the results; Section 7 examines the performance of ABO vis-à-vis Neural Networks methods; Section 8 draws conclusions on the study.

2. Literature Review

Nature-inspired algorithms (NAs) draw their inspiration from the complex but highly organised attributes of nature. Nature-inspired algorithms which are generally stochastic became the hotbed of several researchers due to the inefficiencies of the exact optimization algorithms (as the problem size enlarges) like the Linear Programming, [9], Dynamic Programming [10], finite elements [11], and finite volume methods. In general NAs simulate the interplay and, sometimes, interdependence of natural elements such as plants, animals, and rivers. The most popular class of algorithms among the NAs are the biology-inspired algorithms. A handful of other NAs, however, are inspired by Chemistry or Physics. Some of those algorithms inspired by Chemistry or Physics include Harmony Search (HS) algorithm, Intelligent Water Drop (IWD), Simulated Annealing (SA), and Black Hole (BH) [12]. In this study, our interest is the biology-inspired optimization algorithms.

Biology-inspired algorithms (BAs) can be categorised into three broad classes, namely, Evolutionary Algorithms (EAs) that are inspired by natural evolutions, the Swarm Intelligence (SI) which simulates the collective behavior in plants and animals, and, thirdly, Ecology algorithms which are concerned with simulating the inter- and intracooperative or competitive interactions within the ecosystem [13].

The Evolutionary Algorithms (EAs), generally, simulate the iterative progress comprising growth, development, reproduction, selection, and survival as seen in a population. EAs are concerned with the feasible solutions that emanate iteratively from generation to generation towards the best solution. EAs employ fitness-based selection within the population of solutions in a way that fitter solutions are selected for survival into the next generation of solutions. In this category of EAs are Genetic Programming (GP), Paddy Field Algorithm (PFA), Evolutionary Strategies (ES), Genetic Algorithm (GA), and the Differential Evolution (DE) [12]. It may be necessary to state that experts have different opinions on the classification of Differential Evolution as an Evolutionary Algorithm. The classification of DE as an Evolutionary Algorithm stems from its use of “Evolution” as one of its parameters (just like other EAs). A critical expert may not class Differential Evolution as being bioinspired, actually [14].

The second category of BAs is the Swarm Intelligence (SI) algorithms which are concerned with the collective social behavior of organisms. The motivation of Swarm Intelligence is the collective intelligence of groups of simple agents such as insects, fishes, birds, bacteria, worms, and other animals based on their behavior in real life. As simple as these animals are, they are able to exhibit exceptional intelligence whenever they work collectively as a group. These algorithms track the collective behavior of animals that exhibit decentralized, self-organized patterns in their foraging duties. Examples of these algorithms are the Bee Colony Optimization (BCO), Firefly Algorithm (FFA), Particle Swarm Optimization (PSO), Ant Colony Optimization (ACO), Artificial Bee Colony (ABC), Bacteria Foraging Algorithm (BFA), and so on [15].

The third category of BAs is the Ecology algorithms which are concerned with the numerous inter- or intraspecies competitive or cooperative interactions within the natural ecosystems. The ecosystem is made up of living organisms along with their abiotic environment with which the organisms interact such as water, air, and soil. Cooperation among organisms includes division of labor and represents the core of their sociality. Some of the interactions are cooperative and others are competitive leading to a complex and harmonious balancing within the ecosystem. Algorithms in this category are the PS2O, Invasive Weed Colony Optimization, and biogeography-based optimization [16].

The development of the African Buffalo Optimization (ABO) is in response to the observed lapses in the existing algorithms. The ABO belongs to the Swarm Intelligence class of algorithms based on the social behavior in animals with the aim of achieving greater exploitation and exploration of the search space, ease of use, and faster speed in achieving the optimal result with its use of relatively fewer parameters in solving combinatorial optimization problems.

2.1. Ant Colony Optimization (ACO)

Ant Colony Optimization algorithm is a population-based optimization technique developed by Marco Dorigo and has been successfully applied to solve several NP-hard combinatorial optimization problems. This algorithm was inspired by the behavior of ant colonies, especially, by their foraging behavior in real life. Usually ants, when leaving their nests, move randomly around the areas surrounding their nests in search for food. If any of the ants come across food, it first collects some pieces of the food and, on its way back to the nest, deposits a chemical substance called pheromones as a way of communicating to its peers that there has been a breakthrough. Other nearby ants, on perceiving the fragrance of the pheromone, understand and move towards the pheromone path. Once they discover the food source, they, in turn, drop fresh pheromones as a way of alerting other ants. In a matter of time, several ants pick this information and are on the pheromone path.

Another interesting part of the ants’ behavior is that as they return to the nest, they optimize their route. In a short while, the ants have created a shorter route to the food source than the previous routes. Moreover, in case an obstruction is put on the shorter route, making movements impossible, the ants are able to find another short route among the available options to evade the obstacle. The highlights of this algorithm include tapping into the indirect communication of a colony of (artificial) ants using pheromone trails as a means of communication, tracking their cooperative ability to solve a complex problem, and harnessing their capacity to optimize their routes from the food source to the nest and vice versa.

There have been several modifications of the ant colony algorithms starting from the initial Ant System (AS), to Ant Colony System (ACS), to Min-Max Ant System (MMAS), and then to the Ant Colony Optimization (ACO) algorithms, and so forth [17]. In ACO, a colony of ants in each iteration constructs a solution probabilistically as ant at node selects the next node to move on to. The choice of node is influenced by the pheromone trail value and the available heuristic . In TSP, . So an ant moves from location to location with the probabilityHere, represents the pheromone trail, represents the local heuristic information represents the iteration, represents the nodes ant can go to, and and are parameters that bias the pheromone trails. By the end of an iteration, the pheromone trail on each edge is updated using the following equation:In (3), represents the pheromone trail in iteration ; takes any values from 0.1 to 0.9. The amount of pheromone deposited by the best ant is represented by In (4), represents cost of the best solution .

A critical examination of the Ant Colony Optimization technique of solving optimization problems reveals that there is little or no similarity between the ACO’s search mechanism and that of the ABO. This could be due to their application of different search techniques in arriving at solutions: while the ACO employs path construction technique, the ABO favours path improvement search mechanism.

2.2. Particle Swarm Optimization

Particle Swarm Optimization which was inspired by the social behavior of birds flocking or fish schooling is one of the biology-inspired computation techniques developed by Eberhart and Kennedy [18]. This algorithm obtains solutions using a swarm of particles, where each particle represents a candidate solution. When compared to evolutionary computation paradigms, a swarm is similar to a population and a particle represents an individual. In searching for a solution, the particles are flown through a multidimensional search space, where the position of each particle is adjusted according to its own experience and that of its neighbors. The velocity vector drives the optimization process as well as highlighting each particle’s experience and the information exchange within its neighborhood. Just like the ABO, PSO has two controlling equations in its search for solution and these are (5) and (6). Considerwhere represents the present velocity, is the previous velocity, is the constriction factor, and are the acceleration coefficients, and are random numbers, is the individual particles’ best position, is the present position, and is the swarm’s best position. The next equation in PSO that calculates the position of the swarm is In PSO algorithm, the particles move in the domain of an objective function , where represents the variables to be optimized. Each particle, at a particular iteration, is associated with three vectors:(a)Present position, denoted by : This vector records the present position of a particular particle.(b)Present velocity, denoted by : This vector stores the particle’s direction of search.(c)Individual’s best position, denoted by : This vector records the particular particle’s best solution so far since the search began (since the beginning of the algorithm’s execution). In addition to these, the individual particles relate with the best particle in the swarm which PSO algorithm tracks, in each iteration, to help direct the search to promising areas [19].

2.3. Artificial Bee Colony

This algorithm, which is inspired by the behavior of natural honey bee swarm, was proposed by Karaboga and Akay in 2009 [20]. It searches for solution through the use of three classes of bees: scout bees, onlooker bees, and employed bees. Scout bees are those that fly over the search space in search for solutions (food source). The onlooker bees, on the other hand, are the ones that stay in the nest waiting for the report of the scout bees while the employed bees refer to the class of bees which, after watching the waggle dance of the scout bees, opts to join in harvesting the food source (exploitation). A particular strength of this algorithm lies in its bee transformation capabilities. For instance, a scout bee could transform to an employed bee once it (the same scout bee) is involved in harvesting the food source and vice versa.

Generally, the bees can change their statuses depending on the needs of the algorithm at a particular point in time. In this algorithm, the food source represents a solution to the optimization problem. The volume of nectar in a food source represents the quality (fitness) of the solution. Moreover, each employed bee is supposed to exploit only one food source, meaning that the number of employed bees is the same as the number of food sources. The scout bees are always exploring for new food sources with higher nectar quantity and/or quality (fitness) within the neighbourhood. The bees evaluate the nectar fitness usingwhere is a randomly chosen parameter index; is a random number within a given range; is a food source.

The quality (fitness) of a solution is calculated using the following equation:From the foregoing discussion, it is clear that there is slight similarity between (5) in PSO and (7) in ABO since each algorithm subtracts a variable from the personal and individual bests of the particles/buffalos. For PSO, the subtracted variable is the present position and for ABO, it is the immediate-past explored location (the waaa values, ). However, the two equations are different in several respects: while the PSO uses (being the constriction factor) or (as inertia factor, in some versions of the PSO), there are no such equivalents in ABO. Moreover, while the PSO employs random numbers ( and ), ABO does not use random numbers, only learning parameters. Also, PSO uses acceleration coefficients ( and ); ABO does not. In the case of the ABC, even though it employs the same search technique in arriving at solutions, the algorithm procedures are quite different.

2.4. Information Propagation

In searching for solutions to an optimization problem, the ACO employs the path construction technique while the PSO, ABC, and the ABO use the path improvement technique. However, while the PSO uses the Von Neumann (see Figure 1) as its best technique for information propagation [21], the ACO obtains good results using the ring topology [22] and the ABO uses the star topology which connects all the buffalos together. The Von Neumann topology enables the particles to connect to neighboring particles on the east, west, north, and south. Effectively, a particular particle relates with the other four particles surrounding it. The ABO employs the star topology such that a particular buffalo is connected to every other buffalo in the herd. This enhances ABO’s information dissemination at any particular point in time.

Figure 1: Information propagation topologies.

3. African Buffalo Optimization Algorithm

In using the ABO to proffer solutions in the search space, the buffalos are first initialized within the herd population and are made to search for the global optima by updating their locations as they follow the current best buffalo in the herd. Each buffalo keeps track of its coordinates in the problem space which are associated with the best solution (fitness) it has achieved so far. This value is called representing the best location of the particular buffalo in relation to the optimal solution. The ABO algorithm follows this pattern: at each step, it tracks the dynamic location of each buffalo towards the and depending on where the emphasis is placed at a particular iteration. The speed of each animal is influenced by the learning parameters.

3.1. ABO Algorithm

The ABO algorithm is presented below:(1)Initialization: randomly place buffalos to nodes at the solution space.(2)Update the buffalos fitness values usingwhere and represent the exploration and exploitation moves, respectively, of the th buffalo (); and are learning factors; is the herd’s best fitness and the individual buffalo ’s best found location.(3)Update the location of buffalo ( and ) using (4)Is updating? Yes, go to (5). No, go to (2).(5)If the stopping criteria are not met, go back to algorithm step (3); else go to (6).(6)Output best solution.

A closer look at the algorithm (the ABO algorithm) reveals that (9) which shows the democratic attitude of the buffalos has three parts: the first represents the memory of the buffalos past location. The buffalo has innate memory ability that enables it to tell where it has been before. This is crucial in its search for solutions as it helps it to avoid areas that produced bad results. The memory of each buffalo is a list of solutions that can be used as an alternative for the current local maximum location. The second is concerned with the caring or cooperative part of the buffalos and is a pointer to the buffalo’s social and information-sharing experience and then the third part indicates the intelligence part of the buffalos. So the ABO exploits the memory, caring intelligent capabilities of the buffalos in the democratic equation (9). Similarly, (10) is the waaa vocalization equation that propels the animals to move on to explore other environments as the present area has been fully exploited or is unfavourable for further exploration and exploitation.

3.2. Highlights of the ABO Algorithm

They are as follows:(1)Stagnation handling through regular update of the best buffalo in each iteration.(2)Use of relatively few parameters to ensure speed fast convergence.(3)A very simple algorithm that require less than 100 lines of coding effort in any programming language.(4)Ensuring adequate exploration by tracking the location of best buffalo () and each buffalo’s personal best location ().(5)Ensuring adequate exploitation through tapping into the exploits of other buffalos .

3.3. Initialization

The initialization phase is done by randomly placing the th buffalo in the solution space. For initialization, some known previous knowledge of the problem can help the algorithm to converge in less iterations.

3.4. Update Speed and Location

In each iteration, each buffalo updates its location according to its former maximum location () and some information gathered from the exploits of the neighboring buffalos. This is done using (9) and (10) (refer to the ABO algorithm steps (2) and (3) above). This enables the algorithm to track the movement of the buffalos in relation to the optimal solution.

4. Using ABO to Solve the Travelling Salesman’s Problem

The ABO has the advantage of using very simple steps to solve complex optimization problems such as the TSP. The basic solution steps are as follows:(a)Choose, according to some criterion, a start city for each of the buffalos and randomly locate them in those cities. Consider(b)Update buffalo fitness using (9) and (10), respectively.(c)Determine the and max.(d)Using (11) and heuristic values, probabilistically construct a buffalo tour by adding cities that the buffalos have not visited.(e)Is the updating? Go to (f). No, go to (a).(f)Is the exit criteria reached? Yes, go to (g). No, return to (b).(g)Output the best result.Here, and are learning parameters and are 0.6 and 0.4, respectively, takes the values of 0.5 and −0.5 in alternate iterations, is the positive reinforcement alert invitation which tells the animals to stay to exploit the environment since there are enough pastures, and is the negative reinforcement alarm which tells the animals to keep on exploring the environment since the present location is not productive. For buffalo , the probability of moving from city to city depends on the combination of two values, namely, the desirability of the move, as computed by some heuristic indicating the previous attractiveness of that move and the summative benefit of the move to the herd, indicating how productive it has been in the past to make that particular move. The denominator values represent an indication of the desirability of that move.

4.1. ABO Solution Mechanism for the TSP

The ABO applies the Modified Karp Steele algorithm in its solution of the Travelling Salesman’s Problem [23]. It follows a simple solution step of first constructing a cycle factor of the cheapest weight in the graph. Next, it selects a pair of edges taken from different cycles of the graph and patch in a way that will result in a minimum weight. Patching is simply removing the selected edges in the two cycle factors and then replacing them with cheaper edges and in this way forming a larger cycle factor, thus, reducing the number of cycle factors in graph by one. Thirdly, the second step is repeated until we arrive at a single cycle factor in the entire graph . This technique fits into the path improvement technique description [24]. ABO overcomes the problem of delay in this process through the use of two primary parameters to ensure speed, namely, and , coupled with the algorithm keeping a track of the route of the as well as in each construction step.

5. Experiments and Discussion of Results

In this study, the ABO was implemented on three sets of symmetric TSP datasets and a set of asymmetric TSP (ATSP) ranging from 48 to 14461 cities from TSPLIB95 [25]. The first experiment was concerned with the comparison of the performance of ABO in TSP instances with the results obtained from a recent study [26] involving Berlin52, St70, Eil76, Pr76, KroA100, Eil101, Ch150, and Tsp225. The second set of experiments was concerned with testing ABO’s performance with another recently published study [27] on Att48, St70, Eil76, Pr152, Gil262, Rd400, Pr1002, D1291, Fn14461, and Brd14051. The third experiment examined ABO’s performance in asymmetric TSP instances. The fourth set of experiments involved comparing ABO results with those obtained using some popular Artificial Neural Networks methods [28]. The TSP benchmark datasets are presented in Table 1.

Table 1: TSPLIB datasets.
5.1. Experimental Parameters Setting

For the sets of experiments involving PSO, the parameters are as follows: population size: 200; iteration (): 1000; inertia weight: 0.85; : 2; : 2 . For the HPSACO, the experimental parameters are as follows: population: 200; iteration (): 1000; inertia weight: 0.85; : 2; : 2; ants (N): 100; pheromone factor (): 1.0; heuristic factor (): 2.0; evaporation coefficient (): 0.05; pheromone amount: 100. For the experiments involving other algorithms, in this study, Table 2 contains the details of the parameters used.

Table 2: Experimental parameters.
5.2. Experiment Analysis and Discussion of Results

The experiments were performed using MATLAB on Intel Duo Core i7-3770 CPU, 3.40 ghz with 4 GB RAM. The experiments on the asymmetric Travelling Salesman’s Problems were executed on a Pentium, Duo Core, 1.80 Ghz processor and 2 GB RAM desktop computer. Similarly, the experiments on the Artificial Neural Networks were executed using Microsoft Virtual C++, 2008, on an Intel Duo Core, i7, CPU. To authenticate the veracity of the ABO algorithm in solving the TSP, we initially carried out experiments on eight TSP cases. The city-coordinates data are available in [29]. The results are presented in Table 3.

Table 3: Comparative experimental result.

In Table 3, the “Average Value” refers to the average fitness of each algorithm, and the “relative error” values are obtained by calculatingAs can be seen in Table 3, the ABO outperformed the other algorithms in all test instances under investigation. The ABO, for instance, obtained the optimal solution to Berlin52 and Eil76. No other algorithm did. Besides this, the ABO obtained the nearest-optimal solution to the remaining TSP instances compared to any other algorithm. In terms of the average results obtained by each algorithm, the ABO still has the best performance. It is rather surprising that the Hybrid Algorithm (HA) which uses a similar memory matrix like the ABO could not pose a better result. This is traceable to the use of several parameters since the HA is a combination of the ACO and the ABC: the two algorithms that have some of the largest numbers of parameters to tune in order to obtain good solutions.

The dominance of the ABO can also be seen in the use of computer resources (CPU time) where it is clearly seen that the ABO is the fastest of all four algorithms. In Berlin52, for instance, the ABO is 58,335 times faster than the ACO; 1,085 times faster than the ABC; 30,320 times faster than the Hybrid Algorithm (HA). This trend continues in all the test cases under investigation. To solve all the TSP problems here, it took ABO a cumulative time of 0.279 seconds to ACO’s 7456 seconds; ABC’s 43.11 seconds; and HA’s 3362.27 seconds. The speed of the ABO is traceable to effective memory management technique since the ABO uses the path improvement technique as against the ACO that uses the slower path construction method. The difference in speed with the ABC that uses similar technique with ABO is due to the ABC’s use of several parameters. The speed of the HA could have been affected by the combination of the path construction and path improvement techniques coupled with the use of several parameters.

5.3. ABO and Other Algorithms

Encouraged by the extremely good performance of the ABO in the first experiments, more experiments were performed and the results obtained compared with those from PSO, ACO, and HPSACO from another recently published study [27]. The HPSACO is a combination of three techniques, namely, the Collaborative Strategy, Particle Swarm Optimization, and the Ant Colony Optimization algorithms. The datasets used in this experiment are from the TSPLIB95 and they are Att48, St70, Eil76, Pr152, Gil262, Rd400, Pr1002, D1291, Fn14461, and Brd14051. The results are presented in Table 4.

Table 4: Comparative optimal results.

Table 4 further underscores the excellent performance of ABO. The ABO obtained the optimal solution to Eil76 and Gil262 and very near-optimal solution in the rest of test cases. It is obvious from the relative error calculations that the ABO obtained over 99% accuracy in all the TSP instances here except the difficult rd100 where it obtained 95%. It is worthy of note that the 95% accuracy of the ABO is still superior to the performance of the other comparative algorithms in this TSP instance. The cumulative relative error of the ABO is 5.68% to the PSO’s 61.83%, the ACO’s 39.81%, and the HPSACO’s 28.28%. Clearly, from this analysis, the ABO has a superior search capability. This is traceable to its use of relatively fewer parameters than most other Swarm Intelligence techniques. The controlling parameters of the ABO are just the learning parameters ( and ).

In designing the ABO, the authors deliberately tried to avoid the “Frankenstein phenomena” [30], that is, a case of developing intricate techniques with several parameters/operators that, most times, the individual contributions of some of such parameters/operators to the workings of the algorithm are difficult to pinpoint. The ability to achieve this “lean metaheuristic” design (which is what we tried to do in designing the ABO) is a mark of good algorithm design [30].

6. ABO on Asymmetric TSP Test Cases

Moreover, we had to experiment on some ATSP test cases to verify the performance of the novel algorithm on such cases. The results obtained from this experiment using ABO are compared with the results obtained from solving some benchmark asymmetric TSP instances available in TSPLIB95 using the Randomized Arbitrary Insertion algorithm (RAI) [31], Ant Colony System 3-opt (ACS), Min-Max Ant System (MMAS), and Iterated Local Search (ILS) which is reported to be one of the best methods for solving TSP problems. These ATSP instances are ry48p, ft70, Kro124p, ftv70, p43, and ftv170 [32].

Table 5 shows the comparative results obtained by applying five different algorithms to solving a number of benchmark asymmetric TSP instances with the aim of evaluating the performance of the African Buffalo Optimization (ABO) algorithm. The first column lists the ATSP instances as recorded in the TSPLIB; the second column indicates the number of cities/locations/nodes involved in each instance; the third indicates the optimal solutions, then followed by the performances of the different algorithms in at most 50 test runs.

Table 5: Comparative results.

A quick look at the table shows the excellent performance of the ILS and the MMAS with both obtaining 100% results in the only four cases available in literature that they solved. The ABO performed very well achieving over 99% in all test cases but one. The ABO has approximately the same performance with RAI and the ACS obtaining about 99% optimal results in virtually all cases. It is significant to note, however, that the ABO performed better than ACS and RAI in ft70 which was said to be a difficult ATSP instance to solve in spite of its relatively small size [33].

Next, the authors examined the speed of each algorithm to achieve result since one of the aims of the ABO is to solve the problem of delay in achieving optimal solutions since speed is one of the hallmarks of a good algorithm [34]. In Table 6, we compare the speed of the ABO with those of the recently published Model Induced Max-Min Ant Colony Optimization (MIMM-ACO), Min-Max Ant System (MMAS) Cooperative Genetic Ant Systems (CGAS) [35], and RAI.

Table 6: Comparative speed of algorithms.

Table 6 displays the exceptional capacity of the novel African Buffalo Optimization (ABO) to obtain optimal or near-optimal solutions at record times. It took the ABO less than a second (precisely 0.95 seconds) to solve all the benchmark cases under investigation to MIMM-ACO’s 232.09 seconds; MMAS’s 208.88 seconds; CGAS’ 452.927 seconds; and RAI’s 323.4 seconds. Undoubtedly, the ABO outperformed the other algorithms in its quest to obtain results using very limited CPU time and resources. This brought about ABO’s use of very few parameters coupled with straightforward fitness function. The exceptional speed of the ABO compared with its competitive ability to obtain very competitive results recommends the ABO as an algorithm of choice when speed is a factor.

7. Comparing ABO to Neural Network Solutions to TSP

Moreover, following the popularity and efficiency of Neural Networks in obtaining good solutions to the Travelling Salesman’s Problem [36], the authors compared the performance of African Buffalo Optimization to the known solutions of some popular Neural Network algorithms. These are Angeniol’s method, Somhom et al.’s method, Pasti and Castro’s method, and Masutti and Castro’s method. Comparative experimental figures are obtained from [28].

From the experimental results in Table 7, it can be seen that only the ABO obtained the optimal solutions in Eil51 and Eil76, in addition to jointly obtaining the optimal result with Masutti and Castro’s method in Berlin52. Aside from these, ABO outperformed the other methods in getting near-optimal solutions in Bier127, KroA100, KroB100, KroB100, KroC100, KroD100, KroE100, Ch130, Ch150, KroA150, KroB150, KroA200, KroB150, KroB200 Rat575, rl1323, fl1400, fl1400, and Rat783. It was only in Eil101 that Masutti and Castro’s method outperformed ABO. This is a further evidence that ABO is an excellent performer even when in competition with Neural Networks methods.

Table 7: ABO versus NN results.

8. Conclusion

In general, this paper introduces the novel algorithm, the African Buffalo Optimization, and shows its capacity to solve the Traveling Salesman’s Problem. The performance obtained from using the ABO is weighed against the performance of some popular algorithms such as the Ant Colony Optimization (ACO), Particle Swarm Optimization (PSO), Artificial Bee Colony Optimization (ABO), Min-Max Ant System (MMAS), and Randomized Insertion Algorithm (RAI); some hybrid methods such as Model Induced Max-Min Ant Colony Optimization (MIMM-ACO), Hybrid Algorithm (HA), and Cooperative Genetic Ant Systems (CGAS); and some popular Neural Networks-based optimization methods. In total, 33 TSP datasets cases ranging from 48 to 14461 cities were investigated and the ABO results obtained were compared with results obtained from 11 other optimization algorithms and methods. The results show the amazing performance of the novel algorithm’s capacity to obtain optimal or near-optimal solutions at an incredibly fast rate. The ABO’s outstanding performance is a testimony to the fact that ABO has immense potentials in solving optimization problems using relatively fewer parameters than most other optimization algorithms in literature.

Having observed the robustness of the ABO in solving the Travelling Salesman’s Problems with encouraging outcome, it is recommended that more investigations be carried out to ascertain the veracity or otherwise of this new algorithm in solving other problems such as PID tuning, knapsack problem, vehicle routing, job scheduling, and urban transportation problems.

Conflict of Interests

The authors assert that there is no conflict of interests in the publication of this research paper.


The authors acknowledge the support of the anonymous reviewers for their constructive reviews that helped to enhance the quality of this research paper. Similarly, the authors are grateful to the Faculty of Computer Systems and Software Engineering, Universiti Malaysia Pahang, Kuantan, Malaysia, for providing the conducive environment, facilities, and funds for this research through Grant GRS 1403118.


  1. J. Kennedy, “Particle swarm optimization,” in Encyclopedia of Machine Learning, pp. 760–766, Springer, 2010. View at Google Scholar
  2. M. Dorigo and M. Birattari, “Ant colony optimization,” in Encyclopedia of Machine Learning, pp. 36–39, Springer, 2010. View at Google Scholar
  3. E. Cantú-Paz, “A summary of research on parallel genetic algorithms,” IlliGAL Report 95007, Illinois Genetic Algorithms Laboratory, 1995. View at Google Scholar
  4. D. Karaboga and B. Basturk, “A powerful and efficient algorithm for numerical function optimization: artificial bee colony (ABC) algorithm,” Journal of Global Optimization, vol. 39, no. 3, pp. 459–471, 2007. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  5. J. Andre, P. Siarry, and T. Dognon, “An improvement of the standard genetic algorithm fighting premature convergence in continuous optimization,” Advances in Engineering Software, vol. 32, no. 1, pp. 49–60, 2001. View at Publisher · View at Google Scholar · View at Scopus
  6. N. Kumbharana and G. M. Pandey, “A comparative study of ACO, GA and SA for solving travelling salesman problem,” International Journal of Societal Applications of Computer Science, vol. 2, pp. 224–228, 2013. View at Google Scholar
  7. Z. Michalewicz and D. B. Fogel, “The traveling salesman problem,” in How to Solve It: Modern Heuristics, pp. 189–224, Springer, 2000. View at Google Scholar
  8. J. B. Odili, H. Deng, C. Yang, Y. Sun, M. C. d. E. Santo Ramos, and M. C. G. Ramos, “Application of ant colony optimization to solving the traveling salesman's problem,” Science Journal of Electrical & Electronic Engineering, In press.
  9. A. Ben-Israel, “A Newton-Raphson method for the solution of systems of equations,” Journal of Mathematical Analysis and Applications, vol. 15, pp. 243–252, 1966. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  10. D. P. Bertsekas, Dynamic Programming and Optimal Control, vol. 1, Athena Scientific, Belmont, Mass, USA, 1995.
  11. R. G. Ghanem and P. D. Spanos, Stochastic Finite Elements: A Spectral Approach, Courier Corporation, 2003.
  12. S. Binitha and S. S. Sathya, “A survey of bio inspired optimization algorithms,” International Journal of Soft Computing and Engineering, vol. 2, pp. 137–151, 2012. View at Google Scholar
  13. F. Dressler and O. B. Akan, “A survey on bio-inspired networking,” Computer Networks, vol. 54, no. 6, pp. 881–900, 2010. View at Publisher · View at Google Scholar · View at Scopus
  14. X.-S. Yang and S. Deb, “Two-stage eagle strategy with differential evolution,” International Journal of Bio-Inspired Computation, vol. 4, no. 1, pp. 1–5, 2012. View at Publisher · View at Google Scholar · View at Scopus
  15. S. K. Dhurandher, S. Misra, P. Pruthi, S. Singhal, S. Aggarwal, and I. Woungang, “Using bee algorithm for peer-to-peer file searching in mobile ad hoc networks,” Journal of Network and Computer Applications, vol. 34, no. 5, pp. 1498–1508, 2011. View at Publisher · View at Google Scholar · View at Scopus
  16. A. R. Mehrabian and C. Lucas, “A novel numerical optimization algorithm inspired from weed colonization,” Ecological Informatics, vol. 1, no. 4, pp. 355–366, 2006. View at Publisher · View at Google Scholar · View at Scopus
  17. M. Dorigo, G. Di Caro, and L. M. Gambardella, “Ant algorithms for discrete optimization,” Artificial Life, vol. 5, no. 2, pp. 137–172, 1999. View at Publisher · View at Google Scholar · View at Scopus
  18. M. Clerc, Particle Swarm Optimization, vol. 93, John Wiley & Sons, 2010.
  19. Z. Yuan, M. A. M. de Oca, M. Birattari, and T. Stützle, “Continuous optimization algorithms for tuning real and integer parameters of swarm intelligence algorithms,” Swarm Intelligence, vol. 6, no. 1, pp. 49–75, 2012. View at Publisher · View at Google Scholar · View at Scopus
  20. D. Karaboga and B. Akay, “A survey: algorithms simulating bee swarm intelligence,” Artificial Intelligence Review, vol. 31, no. 1–4, pp. 61–85, 2009. View at Publisher · View at Google Scholar · View at Scopus
  21. J. Kennedy and R. Mendes, “Population structure and particle swarm performance,” in Proceedings of the Congress on Evolutionary Computation (CEC '02), vol. 2, pp. 1671–1676, IEEE, Honolulu, Hawaii, USA, May 2002. View at Publisher · View at Google Scholar
  22. A. M. Mora, P. García-Sánchez, J. J. Merelo, and P. A. Castillo, “Migration study on a pareto-based island model for MOACOs,” in Proceedings of the 15th Genetic and Evolutionary Computation Conference (GECCO '13), pp. 57–64, July 2013. View at Publisher · View at Google Scholar · View at Scopus
  23. M. Dorigo and L. M. Gambardella, “Ant colonies for the travelling salesman problem,” BioSystems, vol. 43, no. 2, pp. 73–81, 1997. View at Publisher · View at Google Scholar · View at Scopus
  24. M. G. C. Resende, R. Martí, M. Gallego, and A. Duarte, “GRASP and path relinking for the max–min diversity problem,” Computers & Operations Research, vol. 37, no. 3, pp. 498–508, 2010. View at Publisher · View at Google Scholar · View at Scopus
  25. G. Reinelt, “Tsplib95, 1995,” 2012,
  26. M. Gündüz, M. S. Kiran, and E. Özceylan, “A hierarchic approach based on swarm intelligence to solve the traveling salesman problem,” Turkish Journal of Electrical Engineering and Computer Sciences, vol. 23, no. 1, pp. 103–117, 2015. View at Publisher · View at Google Scholar · View at Scopus
  27. H. Jia, “A novel hybrid optimization algorithm and its application in solving complex problem,” International Journal of Hybrid Information Technology, vol. 8, pp. 1–10, 2015. View at Google Scholar
  28. S.-M. Chen and C.-Y. Chien, “A new method for solving the traveling salesman problem based on the genetic simulated annealing ant colony system with particle swarm optimization techniques,” in Proceedings of the International Conference on Machine Learning and Cybernetics (ICMLC '10), pp. 2477–2482, July 2010. View at Publisher · View at Google Scholar · View at Scopus
  29. S. Georg, “MP-TESTDATA—The TSPLIB symmetric traveling salesman problem instances,” 2008.
  30. K. Sörensen, “Metaheuristics—the metaphor exposed,” International Transactions in Operational Research, vol. 22, no. 1, pp. 3–18, 2015. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  31. J. Brest and J. Zerovnik, “A heuristic for the asymmetric traveling salesman problem,” in Proceedings of the 6th Metaheuristics International Conference (MIC '05), pp. 145–150, Vienna, Austria, 2005.
  32. M. Dorigo and L. M. Gambardella, “Ant colony system: a cooperative learning approach to the traveling salesman problem,” IEEE Transactions on Evolutionary Computation, vol. 1, no. 1, pp. 53–66, 1997. View at Publisher · View at Google Scholar · View at Scopus
  33. O. C. Martin and S. W. Otto, “Combining simulated annealing with local search heuristics,” Annals of Operations Research, vol. 63, pp. 57–75, 1996. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at Scopus
  34. B. Baritompa and E. M. Hendrix, “On the investigation of stochastic global optimization algorithms,” Journal of Global Optimization, vol. 31, no. 4, pp. 567–578, 2005. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  35. J. Bai, G.-K. Yang, Y.-W. Chen, L.-S. Hu, and C.-C. Pan, “A model induced max-min ant colony optimization for asymmetric traveling salesman problem,” Applied Soft Computing, vol. 13, no. 3, pp. 1365–1375, 2013. View at Publisher · View at Google Scholar · View at Scopus
  36. N. Aras, B. J. Oommen, and I. K. Altinel, “The Kohonen network incorporating explicit statistics and its application to the travelling salesman problem,” Neural Networks, vol. 12, no. 9, pp. 1273–1284, 1999. View at Publisher · View at Google Scholar · View at Scopus