Abstract

Atom search optimization algorithm has good searching ability and has been successfully applied to calculate hydrogeological parameters and groundwater dispersion coefficient. Since the atom search optimization algorithm is only based on the atom force motion model in molecular dynamics, it has some shortcomings such as slow search speed and low precision during the later stage of iteration. A modified atom search optimization based on the immunologic mechanism and reinforcement learning is proposed to overcome the abovementioned shortcomings in this paper. The proposed algorithm introduces a vaccine operator to better utilize the dominant position in the current atom population so that the speed, accuracy, and domain search ability of the atom search optimization algorithm can be strengthened. The reinforcement learning operator is applied to dynamically adjust the vaccination probability to balance the global exploration ability and local exploitation ability. The test results of 21 benchmark functions confirm that the performance of the proposed algorithm is superior to seven contrast algorithms in search accuracy, convergence speed, and robustness. The proposed algorithm is used to optimize the permutation flow shop scheduling problem. The experimental results indicate that the proposed algorithm can achieve better optimization results than the seven comparative algorithms, so the proposed algorithm has good practical application value.

1. Introduction

Optimization has been a hot topic in scientific research and engineering application [1] such as industry, society, economy, and management. The optimization methods include traditional exact solution method, constructive algorithm, and swarm intelligence algorithm [2]. The traditional exact solution method can obtain the exact solution, but it requires the objective function to be continuous and differentiable. The algorithm of traditional exact solution method is too complex, so it is suitable for solving small-scale problems. Meanwhile, the traditional exact solution method is not good at optimizing multipeak, nonlinear, and dynamic problems. The constructive algorithm [3] can quickly obtain the solution, but the solution quality is poor and difficult to meet the actual engineering needs. The swarm intelligence algorithm simulates the swarm behavior of social organisms to optimize the given objectives and embodies the characteristics of randomness, parallel, and distribution. The swarm intelligence algorithm provides a solution to complex problems without centralized control and without providing the global mathematical model. Compared with traditional optimization methods, the swarm intelligence algorithm has simple principle and fewer parameters and does not need gradient information of the problem. At present, the common swarm intelligence algorithms include particle swarm optimization algorithm [4], differential evolution algorithm [5], cuckoo search algorithm [6], flower pollination algorithm [7], sine-cosine search algorithm [810], and artificial bee colony algorithm [11]. They are widely used in function optimization [1214], combinatorial optimization [15], image segmentation [16], and flow shop scheduling [1719].

The atom search optimization (ASO) was proposed by Zhao et al. in 2018 based on the force motion model of atoms in molecular dynamics [20]. ASO is a type of physics-inspired swarm intelligence algorithm. The algorithm needs fewer parameters and has good global exploration ability in the optimization process, so it has been successfully applied to calculate hydrogeological parameters and groundwater dispersion coefficient [21, 22]. Experimental results have proved that the ASO is significantly better than particle swarm optimization and genetic algorithm [20]. Since ASO is proposed only based on the forced motion model of atoms in molecular dynamics, it has the shortcomings of low search accuracy and slow convergence speed as other swarm intelligence algorithms.

The particles distribute approximately evenly in nature and constitute a complex group. The individuals of particles exchange locally acquired information and interact in group. Inspired by the above facts, the modified atom search optimization algorithm based on immunologic mechanism and reinforcement learning (MASO) is proposed to enhance the ASO. The immunologic mechanism can make better use of the dominant positions in the current atomic population to promote the speed and accuracy of the MASO. The reinforcement learning is applied to dynamically adjust the vaccination probability to balance the global exploration ability and local exploitation ability of the MASO.

2. Atom Search Optimization Algorithm

Matter is composed of molecules. Atoms are connected by covalent bonds to form molecules. Atoms have mass and volume. The interaction forces between atoms are shown as repulsions or attractions according to the different distances between atoms. In the repulsive region, the repulsion forces between atoms increase sharply as the distances decrease. In the attractive region, when the distances increase to a certain extent, the attraction forces reach maxima. As the distances continue to increase, the attraction forces gradually decrease to zero. When two atoms are at an equilibrium distance, the interaction force between them is zero. In molecules, atomic forces also need to consider the effects of geometric constraints and atoms’ internal motions. The geometric constraints and atoms’ internal motions are called binding force. Therefore, the constrained atomic motion equation can be simply written as follows:where is the resultant force of the interaction on atom i, is the resultant force of the constraining force on atom i, is the mass, and is the acceleration of atom i.

In ASO, each atom represents a feasible solution in the search space. The mass of an atom represents the quality of a feasible solution. The better the solution is, the higher the mass of the atom will be, and vice versa. All atoms in a population attract or repel each other according to their distances, which cause lighter atoms to move toward heavier atoms. Heavier atoms have less accelerations, which allow them to exploit local space better. Lighter atoms have more accelerations, which allow them to explore a wider search space. The mass of the atom i at the iteration t can be simply measured by the fitness function. The expression of is shown in equation (3):where N is the total number of atoms, is the fitness function value of the atom i at the iteration t, and and are the fitness function values of the worst and best atoms at iteration t, respectively.

In ASO, Kbest is defined as a set of the first K atoms that have the best fitness values in the atomic population. The set is also called the K-nearest neighbors of atom i. The atom i needs to interact with its K-nearest neighbors as much as possible to strengthen exploration at the early iteration stage. The atom i needs to interact with its K neighbor as little as possible to enhance the exploitation at the late iteration stage. Therefore, K decreases as the number of iterations increases. When the total number of iterations is T, K can be calculated as follows:

Since ASO needs to increase the attraction force and reduce the repulsive force with the increase of iteration times, Zhao et al. [20] proposed a simplified model of the interaction . The expression of is as follows:where is a depth function used to adjust the repulsion region and the attraction region. is defined in equation (6), where is the depth weight. According to equation (7), the distance function is updated, where and are the upper and lower limits of h, respectively. is the Euclidean distance between atom i and atom j in the iteration t. is the length, which is defined as follows:where is the position component of atom i in the dimension j of the search space. The dynamically changed and are defined in the following equation:where is the drift component and is defined as follows:

The resultant force of the interaction force on atom i is expressed in the following equation, where is a random number that obeys the uniform distribution of 0-1:

Geometric constraints in molecular dynamics theory play an important role in atomic motion. Suppose that every atom has covalent bonds with every atom in Kbest, every atom is constrained by Kbest. The constraining force of the atom i can be described by the following equation:where is the Lagrange multiplier, is the multiplier weight, and is the position component in dimension d of the best atom at iteration t. According to equations (1), (5), and (12), the updated equation of acceleration can be deduced as follows:

According to the above equations, the velocity and position update equations of atom i at iteration t + 1 are equations (15) and (16), respectively:

The pseudocode of ASO is given in Algorithm 1.

(1)Randomly initialize a set of atoms X (solutions) and their velocity , and .
(2)While the stop criterion is not satisfied do
(3) For each atom do
(4)  Calculate the fitness value Fit;
(5)  If then
(6)   ;
(7)   ;
(8)  End If.
(9)  Calculate the massing using equations (2) and (3);
(10)  Determine its K neighbors using equation (4);
(11)  Calculate the interaction force and the constraint force using equations (11) and (12), respectively;
(12)  Calculate the acceleration using equation (14);
(13)  Update the velocity using equation (15);
(14)  Update the position using equation (16);
(15) End For.
(16)End While.
(17)Find the best solution so far

3. Modified Atom Search Optimization Based on Immunologic Mechanism and Reinforcement Learning

3.1. Chaos

Chaos is a nonlinear phenomenon widely existing in nature. It has the characteristics of randomness, ergodicity and inherent regularity [2325]. Randomness means that the chaotic behavior has the disorderly features similar to random variable. Ergodicity means that chaos can traverse all states according to its own laws without repetition. Regularity means that chaos is generated by deterministic iteration. Therefore, the chaos strategy is used to initialize the swarm intelligence algorithm and avoid the algorithm falling into the local optimum.

There are defects in random initialization of the ASO. These defects can be well solved by the chaotic strategy. In MASO, nine different chaotic maps are used to initialize the atomic population. These nine kinds of mapping will be called circularly to ensure that different problems can be properly initialized. Table 1 displays the basic information about the nine maps. In this paper, the chaos strategy used to initialize population is called the chaos initialization operator.

3.2. Artificial Immune Algorithm and Vaccine Operator

Artificial immune system is an intelligent method inspired by the biological immune system. It is a new information processing system based on the theory of the human immune system. The artificial immune system provides evolutionary learning mechanisms such as noise tolerance, nonteacher learning, no negative examples, and self-organization. The artificial immune system provides a new way to solve the optimization problem of high dimension [26, 27].

The artificial immune algorithm is essentially a group of operators. The artificial immune algorithm can generate and maintain the diversity of population and has the ability of self-regulation, so it is suitable for optimizing and improving the swarm intelligent algorithm.

The ASO does not make full use of the dominant position in the population, so the accuracy and speed of convergence are not outstanding enough. These defects can be improved by introducing the vaccine operator [2831]. The vaccine operator consists of three parts: vaccine extraction operator, vaccination operator, and immune detection operator.(1)Vaccine Extraction Operator. In the artificial immune algorithm, vaccine is the prior knowledge of the solving problem. The vaccine extraction operator uses the iteration results of ASO as the prior knowledge and extracts the best current position of atoms as vaccine. The formula of vaccine extraction operator is shown below: expressed the fitness function value of in the t iteration, and represents the fitness function value of the best position obtained by the iteration so far.(2)Vaccination Operator. Vaccination uses prior knowledge to locally adjust the solutions so that the qualities of the candidate solutions are significantly improved. The vaccination operator will be used to inoculate atomic population. Firstly, a part of the individuals is selected from the atomic population, and then the position of the selected atomic individuals is modified by the following equation:where is the atomic population after vaccination. r is a random number that obeys the uniform distribution of 0-1.(3)Immunoassay Operator. After the operation of the vaccination operator, the immune detection operator compares the fitness values of the vaccinated individuals and the original individuals. If the fitness values of vaccinated individuals are not as good as the original individuals, it means that there is degradation in the process of inoculation. Then the original individuals become the parents of the next generation directly according to equation (19). Otherwise, the vaccinated individuals become the parents of next generation according to the following equation:

The process of using the immune mechanism to enhance ASO is shown in Figure 1.

3.3. Reinforcement Learning Mechanism Updates Vaccination Probability
3.3.1. Reinforcement Learning

Machine learning includes supervised learning, unsupervised learning, and reinforcement learning. Reinforcement learning takes environmental feedback as input. This method is different from the supervised learning and unsupervised learning. The supervised learning and unsupervised learning tell the Agent what behavior to take through positive and negative examples, but reinforcement learning finds the optimal behavior strategy through trial and error [32].

The principle of reinforcement learning is shown in Figure 2. Reinforcement learning regards learning as a test evaluation process. The Agent chooses an action for the environment, and the environment produces a reinforcement signal (award or punish) feedback to the Agent after its state changes. Then the Agent selects the next action according to the reinforcement signal and the current state of the environment. The principle of selection is to increase the probability of receiving positive reinforcement (award). The action of selection affects not only the immediate reinforcement value but also the state of the environment at the next moment and the final reinforcement value [33, 34].

3.3.2. Adjust Vaccination Probability Dynamically

In the MASO, the vaccination probability Pv is determined according to the actual situation. The reinforcement learning method is used to determine the vaccination probability dynamically. In this paper, this method is collectively called the reinforcement learning operator.

The reinforcement learning method used here is based on snap-drift neural network [3537]. The reinforcement learning method switches between snap mode and drift mode [38]. In the snap mode, as Pv decreases, global exploration increases, thus avoiding premature convergence. In the drift mode, as Pv decreases, local exploitation increases, thus improving convergence speed. The Pv value is updated with equation (20). Agent (MASO) accepts the state (snap or drift) and reward value (Pm) at time t and then takes an action (increase Pv or decrease Pv) to convert to a new state. The Pm value is updated with the following equation:where Se is the number of atomic individuals updated in this iteration. Pm is the conversion probability. The value of Pm is determined by the ratio of updated individuals in the population. When Pm is less than 50%, the snap mode is used; when Pm is more than 50%, the drive mode is used. is the step size of Pv for each change. Figure 3 illustrates how reinforcement learning operator works.

3.4. Algorithm Flow

After summarizing the above improved methods of ASO, the pseudocode of MASO is given in Algorithm 2.

(1)Chaotic initialize a set of atoms X (solutions) and randomly initialize their velocity , evaluate the fitness values by using the objective function, and then calculate and . Determine the parameters.
(2)While the stop criterion is not satisfied do
(3) For each atom do
(4)  Calculate the massing using equations (2) and (3);
(5)  Determine its K neighbors using equation (4);
(6)  Calculate the interaction force and the constraint force using equations (11) and (12), respectively;
(7)  Calculate the acceleration using equation (14);
(8)  Update the velocity using equation (15);
(9)  Update the position using equation (16);
(10)  Check out, update the fitness values;
(11)  Invoke the reinforcement learning operator by equations (20) and (21);
(12)  Invoke the vaccination operator by equations (17) and (18);
(13)  Invoke the immune detection operator by equation (19);
(14) End For.
(15)End While
(16)Find the best solution so far

4. Simulation Experiment

The performance of MASO will be verified by experiments in three aspects: (1) The MASO will compare with seven swarm intelligence algorithms using 21 benchmark functions [3941] of 30 dimensions. (2) The significance of MASO improvement will be analyzed by the T test. (3) The significance of MASO improved by individual strategies will be analyzed with the Wilcoxon rank sum test.

4.1. Test Platform and Benchmark Functions
4.1.1. Test Platform

The server model is the dawning 5000 A server. The server configures Xeon X5620 CPU (4 cores) 2, 24 GB memory, and 300 GB SAS hard disk. The server is equipped with Rhel5.6 operating system. The programming tool is MATLAB 2012a (for Linux).

4.1.2. Benchmark Functions

21 benchmark functions are selected to evaluate the algorithm performance. This set of benchmark functions has widely been used to evaluate the ability of the swarm intelligence algorithm [38]. The number, range, type, and formulation of functions are listed in Table 2. The 21 benchmark functions can be divided into three categories. F1∼F6 are the unimodal functions (U) and are applied to investigate the search convergence accuracy. F7–F17 are the complex multimodal functions (M) with multiple local extremum points. These functions are employed to test the global search capability and the ability to avoid prematurity. F18–F21 are the rotating multimodal functions (R). These functions have more extreme points, thus increasing the difficulty of searching. F18–F21 can be adopted to further detect the overall performance of the algorithm.

4.2. Experimental Results Analysis of MASO and Comparison Algorithms

In this experiment, the parameters setting of the comparison algorithms are displayed in Table 3. For each benchmark function, each algorithm runs 20 times independently. The population number is 100. The evaluation times of MASO and CS are  100 in each iteration, and evaluation times of other algorithms are 100 in each iteration. For each benchmark function, MASO and CS will iterate 2500 times and other algorithms will iterate 5000 times. So all algorithms will be compared under the evaluations of  5000 in each run. The various performances of algorithms are measured by the four indexes of optimal value, average value, worst value, and standard deviation. The experimental statistical results are shown in Table 4. Take 500 points to draw the convergence curve. MASO and CS take one point every 50 iterations, and other algorithms take one point every 100 iterations. Figures 424 are the convergence curves of each function successively.

The capacities of MASO in three kinds of functions will be analyzed based on the results of Table 4 and Figures 424.

The MASO has a strong overall performance for single mode functions F1–F6. The single mode functions are relatively smooth, with only one extreme point at the origin, so the MASO is easy to converge to global optimum. In F1, F2, F3, and F5, the minimum, average, maximum, and variance of MASO are the best among the all algorithms. In F4, except BAT, CS, and PSO, the minimum, average, maximum, and variance of other algorithms get the optimums. In F6, only MASO converges to the best minimum compared with other algorithms.

In complex multimodal functions F7–F17, MASO generally performs better among the all algorithms. In F9, F10, F12, F13, F14, and F17, the minimum, average, maximum, and variance of MASO are the best among the all algorithms. In F7, F11, and F16, only MASO converges to the global optimum, compared with other algorithms. In F14, the performance of MASO is worse than ASO.

In the rotated multimodal functions F18–F21, the performance of MASO is not significantly superior to other comparison algorithms. There are too many extreme points in the rotating multimodal function, so MASO needs a strong global exploration ability to converge to the global optimum. On the contrary, MASO needs to consider both local development and global search, so it cannot unilaterally maximize the global search capability. This will affect the performance of MASO in optimizing rotating multimodal functions. In F18, F19, and F21, only MASO converges to the global optimum compared with other algorithms. In F20, MASO and ASO converge to the global optimum. The stability of MASO is the best in F19, but in F18, F20, and F21, the stability of MASO is not good enough.

4.3. Effectiveness Analysis of Improved Strategies

ASO is improved in the following three aspects to obtain MASO: (1) The atomic population is initialized by the chaotic operator to make the population location more uniform so that the MASO does not fall into local optimum easily. (2) The dominant positions in the population are used more fully by the vaccine operator to accelerate the MASO convergence. (3) The vaccination probability Pv will be determined dynamically by the reinforcement learning mechanism so that Pv adapts the problem automatically in a more scientific and reasonable way. The MASO improved from the above three aspects has high search accuracy and convergence speed. The improvement effectiveness of MASO will be analyzed by the T test and rank sum test, respectively.

4.3.1. The T Test Analysis between MASO and Comparative Algorithms

Table 5 lists the results of T test, where “+” means that MASO has obvious advantages when compared with other algorithms, “=” indicates that the T test results of MASO equals other algorithms, and “≈” means that the T test results of MASO are approximately equal to other algorithms.

In F5, F7, F9, F10, F11, F12, F13, F14, F18, F19, and F21, MASO is more prominent than the most comparative algorithms. In F1, F3, F4, F6, F8, F15, F17, and F20, although performance of MASO is not outstanding, but its advantages are still better than most comparative algorithms. In F2 and F16, the effect of MASO is almost the same as the corresponding comparison algorithm. The improvement of MASO is not significant enough in F2 and F16 because most algorithms can optimize the two functions easily. Except F2 and F16, the effect of MASO is obviously better than most other algorithms, so the improvement of MASO is evidently effective. Therefore, MASO is a successful improvement to ASO.

4.3.2. Wilcoxon Rank Sum Test for Improved Strategies

MASO is obtained by using the chaotic initialization operator, vaccine operator, and reinforcement learning operator to ameliorate ASO. The ASO enhanced by the chaotic initialization operator is named ASO_chaos. The ASO enhanced by the vaccine operator is named ASO_vaccine. The ASO enhanced by the reinforcement learning operator is named ASO_rl.

In this experiment, each algorithm runs 20 times independently in each benchmark function. The population number is 100. The evaluation times of ASO_vaccine and ASO_rl are  100 in each iteration, and evaluation times of ASO and ASO_chaos are 100 in each iteration. For each benchmark function, ASO_vaccine and ASO_rl will iterate 2500 times, and ASO and ASO_chaos will iterate 5000 times. So the four algorithms will be compared under the evaluation of  5000 in each run.

Table 6 lists the statistical results with different strategies. Among the three improved methods, the improvement effect of ASO_chaos is relatively general, and its stability is not good too. ASO_vaccine and ASO_rl are slightly improved compared with the ASO, but the improvement effect is not significant. Generally, the minimums of ASO_vaccine and ASO_rl are at most several orders of magnitude lower than the minimum of ASO. Only the MASO algorithm, that is, enhanced by three operators, can achieve remarkable results.

Table 7 displays the statistical results of the Wilcoxon rank sum test, where “+” means that MASO has obvious advantages when compared with other algorithms and “≈” means that the Wilcoxon rank sum test results of MASO approximately equal to other algorithms. In F5, F9, F10, F13, F14, F19, and F21, ASO_chaos, ASO_vaccine, and ASO_rl are more effective than the ASO. The ASO_chaos has better performance than ASO in F15. The improved effect of ASO_chaos, ASO_vaccine, and ASO_rl is not significant in functions F1–F4, F6–F8, F11–F12, F16–F18, and F20. The effect of MASO is significantly better than ASO and comparative algorithms in most benchmark functions. Therefore, the remarkable effect of MASO can only be achieved by using three operators together.

5. Application of MASO in Optimization of Permutation Flow Shop Scheduling Problem

In order to verify the optimization capability of MASO for practical engineering problems, MASO will be used to optimize the permutation flow shop scheduling problem.

5.1. Permutation Flow Shop Scheduling Problem

The flow shop scheduling problem studies the process of n jobs on m machines. According to different constraints, flow shop scheduling can be divided into the following different problems: permutation flow shop scheduling problem, no-wait flow shop scheduling problem, no-idle flow shop scheduling problem, blocking flow shop scheduling problem, flow shop scheduling problem with limited buffers, lot-streaming flow shop scheduling problem, and hybrid flow shop scheduling problem [42].

The solution of the permutation flow shop scheduling problem is to find a scheduling sequence so that a certain production index can reach the optimal value. The optimization objective is the minimum makespan. The minimum makespan equals to the completion time of the last workpiece on the last machine.

5.2. Model Assumptions

Some assumptions are made to establish the mathematical model of permutation flow shop scheduling problem:(1)N workpieces are processed on M machines. Each workpiece must be processed in sequence on different machines without interruption.(2)There is an infinite buffer area between machines.(3)In each process, a workpiece can only be processed on one machine without interruption and a machine can only process one workpiece at the same time.(4)The processing time of the workpiece is known, and the preparation time of the workpiece and the start time of the machine are all included in the processing time.

The mathematical models of permutation flow shop scheduling problems are shown in equations (22), (23), and (24) [43].(1)Objective function:(2)Constraint conditions:

The number of workpieces is n, and a single workpiece is marked as j (j = 1, …, n). The number of machines is s, and a single machine is marked k (k = 1, …, s). is the start time of workpiece j in machine k. is the processing time of workpiece j on machine k. is the makespan of workpiece j in machine k.

5.3. Engineering Sample

An engineering example of permutation flow shop scheduling problem is given below. Suppose three jobs A, B, and C are processed on three machines in the order of A, B, and C, respectively. The starting time of the first job on the first machine is 0. The processing time of all jobs on each machine is listed in Table 8. The Gantt chart and the makespan are shown in Figure 25.

5.4. Experimental Design

The configuration data about the size and structure of the problem need to be obtained to systematically evaluate the performance of algorithm. The actual data in industry are difficult to obtain, so the test data will be generated randomly. These test data will represent the industrial data. The test is divided into nine problem combinations. Each problem can be identified in the form of “number of jobs-number of machines.” For example, problems with 20 jobs and 4 machines can be expressed as “20-4.” The factors of level and scope are given in Table 9.

The setting of experimental parameters is the same as in Section 3.2. But for each test problem, MASO is compared with ASO, FPA, BAT, PSO, CS, ACOR, and CSA by running ten times separately.

5.5. Analysis of Experimental Results

The above eight algorithms have been tested ten times on nine combinations of problems, and the test results are listed in Table 10. The minimum makespan and average are two important performance evaluation indexes. The minimum makespan represents the optimization ability of the algorithm. The average represents the stability of the algorithm.

In the “10-2” problem, the minimum makespan obtained by MASO is better than BAT, PSO, and ACOR and is equal to FPA and CS but is worse than the ASO. The stability of MASO is in the middle of all the test algorithms. Generally, the effect is not ideal. In “10-4,” the search effect of MASO is general, but the average is the best, and the stability is good. In the “10-8” problem, the search ability of MASO is the best, but it is equal to CSA, ACOR, and BAT, respectively. And the average of MASO is the best in all algorithms. In summary, MASO has almost the same performance as other algorithms under the problem scale of 10 jobs because the scale of the problem is too small to reflect the good search capability of MASO.

In the “20-2” problem, MASO has the better performance. The minimum makespan of MASO is equal to CSA and is slightly inferior to PSO. In the “20-4” problem, the minimum makespan of MASO is equal to ACOR and CSA and is slightly inferior to FPA. The minimum makespan of MASO ranks second among all the algorithms. In the “20-8” problem, the optimization capability of MASO is only inferior to BAT, ranking second among all algorithms, but its average is the best. In summary, the overall performance of MASO is better than all other algorithms under the scale of 20 workpieces because MASO has strong optimization accuracy and can coordinate the balance between global exploration and local development. In addition to MASO, ACOR performs well on the scale of the problem, so ACOR is also suitable for optimizing this complex permutation flow shop scheduling problem.

In the problem of “50-2,” the search effect of MASO is the best, but the average of MASO is worse. In the “50-4” problem, the search effect of MASO is only slightly worse than ACOR. In the “50-8” problem, the search ability of MASO is the best and far better than other algorithms.

Therefore, MASO works better on the problem of higher complexity, especially on the problem of 20 and 50 workpieces. The reason is that the immune operator and reinforcement learning operator strengthen the utilization of the dominant position in the atom population, so the search speed and accuracy of MASO are promoted. Experiments confirm that MASO has advantages to optimize engineering problem.

6. Conclusion

The MASO is proposed to improve ASO in this paper. The chaotic operator, immune operator, and reinforcement learning operator are introduced to enhance ASO. The chaotic operator overcomes the disadvantages of nonuniform initialization in ASO. The immune operator and reinforcement learning operator can make better use of the dominant position in the atom population and employs adaptive parameters to avoid human interference factors. The experimental results in 21 benchmark functions indicate that MASO is better than the seven comparison algorithms in global search ability and convergence speed. T test proves that the performance of MASO is significantly better than the seven comparison algorithms on most benchmark functions. The Wilcoxon rank sum test illustrates that the excellent performance of MASO is obtained by the interaction of three operators. When MASO is applied to optimize the permutation flow shop scheduling problem, the search accuracy of MASO is better than the seven contrast algorithms. MASO does not require the objective function to be convex, continuous, or derivable, so it has strong advantages in the field of numerical optimization. Therefore, MASO can be applied in machine learning, engineering optimization, municipal traffic management, and other optimization fields.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

Acknowledgments

This work was supported by the Guangxi Key Laboratory of Multimedia Communications and Network Technology. This work was financially supported by the Natural Science Foundation of China (Grant no. 61761006) and the Natural Science Foundation of Guangxi Province (Grant no. 2017GXNSFAA198263).