# Introduction he fundamental function of a system is to provide its customers with a fairly economical cost, acceptable reliability required. These constraints require an optimal design. In the engineering context, the fundamental interest of manufacturers is to find a balance between the reliability of a system and its cost. These two factors constitute the most important decision variable for optimizing a system. This is generally manifested in minimizing the cost under the constraint of reliability on the one hand, and in improving performance to meet the needs of customers under the constraint of cost on the other. In this area, researchers have developed and improved many methods and algorithms. The set of methods can be divided into two main categories: exact methods, which guarantee to obtain an optimal solution for problems of reasonable size, and approximate methods (heuristics and meta-heuristics), which give good solutions. Quality, without guarantee of optimality, but for the benefit of shorter calculation time. If the exact methods are based on the enumeration, often implicitly, for the search for the set of solutions of the search space, then the approached methods rather require random processes in the exploration of potential solutions, and this to deal with the combinatorial explosion generated when using the exact methods. In this perspective, we are essentially seeking in this work to integrate an efficient and adapted optimization method to solve the problem of optimizing reliability by taking into account the most relevant constraints. Given the large number of configurations that can result from a set of components, the procedure of enumerating all the possible architectures is no longer pleasant. Consequently, it is necessary to opt for an approximate method, which will make it possible to find the solution closest to the optimal solution because it is not obvious to examine all the possibilities. In this perspective, we opted for the method of genetic algorithms as an optimization technique. It is considered an effective meta-heuristic method in the field of dependability. It is inspired by genetic biology and is based on the principle of the search for evolution. It does not guarantee an exact solution, but it generates a solution close to the optimum (Painton and Campbell [1], Levitin and Lisnianski [2]). # II. The Functioning of the Genetic Algorithm a) Origin and principle Genetic algorithms (GA) are heuristic optimization algorithms based on the principles of natural selection and genetics. The researcher Rechenberg [3], is the first scientist who introduced evolutionary algorithms by publishing his work ''Evolution strategies.'' These algorithms are broadly inspired by Darwin's theory of evolution published in 1859. Next, Holland [4] proposed the first genetic algorithms to solve combinatorial optimization problems, and they were also developed by the work of David Goldberg published in 1989 (Goldberg [5] and Goldberg [6]). The aim of the genetic algorithm is to bring up, from one generation to another, the candidates (potential solutions) most suited to solving the problem. Each generation is made up of a defined number of individuals, these form a population, and each of them represents a point in the search space. Each individual (chromosome) has information coded in the form of a Ngnassi Djami Aslain Brisco ? , Nzie Wolfgang ? & Doka Yamigno Serge ? chain of characters that analogically constitutes genes. Then the passage from one generation to another is carried out based on the process of evolution by the use of evolutionary operators like selection, crossing, and mutation. Their operating principle is quite simple. From an initial population created at random, composed of a set of individuals (chromosomes), we proceed to the evaluation of their parent qualifications to highlight the best suited, as long as the least effective are rejected. Then, the most qualified individuals are chosen by privileged selection by giving them a chance to reproduce by crossing and mutating via the two operators of crossing and mutation. Then by relaunching this process several times, the optimal solution can be refined by passing from one generation to another (Douiri et al. [7]). # b) Description of the formalism used Clustering is a process that partitions a set of data into meaningful subclasses (clusters or clusters). The convergence of genetic algorithms has been demonstrated for many problems, although optimality cannot be guaranteed. The ability of a genetic approach to find the right solution often depends on the adequacy of the coding, the evolution operators, and the measures of adaptation to the problem being addressed. The method proposed here is based on genetic algorithms (Goldberg [6]) and evolutionary strategies (Schewefel [8]). It combines the principle of survival of the ablest individuals and genetic combinations for an elitist research mechanism. The genetic method produces new solutions (children) by combining existing solutions (parents) selected from the population, or by mutation. The central idea is that parent solutions will tend to produce superior child solutions in terms of adaptation so that ultimately a solution obtained is optimal. In this study, we used a genetic method previously defined by Bicking et al. [10] with a definition of the chromosome and the operators of selection, combination, and mutation concerned. Unlike genetic algorithms, the genetic method used is designed to minimize and not maximize. This method, like genetic algorithms, is not limited by assumptions about the objective function and research space, such as continuity or differentiability. It uses a population of points simultaneously by contrast with usual methods using only one point. Genetic operators are elitistically improving the search process to find the global optimum. There are more complicated genetic operators, but the basic operators and their various modifications can generally be applied. The choice of these operators depends on the nature of the problem and the performance requirements. The genetic algorithm that we are going to implement is as follows, where the process is applied to iteration k: We will choose, as a stop test in our implementation, a finite number of iterations. It is important to note that the stopping criterion can be several cycles of the algorithm (number of generations), the average of the adaptations of individuals, a convergence factor, etc. An individual represents a vector of decision variable (parameters), and its adaptation is measured by the objective function. The formalism and the genetic operators are detailed below. # i. Data coding The first step is to properly define and code the problem. That step associates with each point of the search space a specific data structure called a chromosome, which will characterize each individual in the population. This step is considered to be the most important step in GA because the success of these algorithms depends heavily on how individuals are coded. There are different choices for coding a chromosome, this choice being a very important factor in the progress of the algorithm so it must be well suited to the problem being addressed: ? Binary coding: It is the most used coding. The chromosome is coded by a string of bits (which can take the value 0 or 1) containing all the information necessary to describe a point in space; ? Multi-character coding: this is often more natural. We are talking about multiple characters as opposed to bits. A chromosome is then represented by a series of numbers or characters, each representing a gene; ? Coding in the form of a tree: this coding in tree structure starts from a root (comprising several parts equal to the number of initial individuals), from which one or more children can be derived. The tree then builds up gradually, adding branches to each new generation. ii. Generation of the initial population Each chromosome is the potential result of the optimization problem. We define a chromosome as a chain composed of genes, which are the parameters (decision variables) to find. The value of a gene is called an allele. The possible value of an allele is an integer or a real value. Each gene is created randomly, using equation 1. (1) Where: ? j ?{0;1} is chosen randomly are the minimum and maximum limits of the allele a j . They are chosen according to the problem to be treated. Each chromosome, called an individual in a haploid representation, can be written: With: m is the number of genes i = 1,..., N and N is the size of the population (number of individuals). All the constraints are taken into account in the initial phase of population creation. When an individual is created, if the constraints are respected, this individual is integrated into the initial population; otherwise, it is not. At the start of the algorithm, the initial population contains individuals. The length of the chromosome m and the size of the population N is two of the four adjustment parameters of the genetic method. # iii. Objective function and adaptation We evaluate the different solutions proposed to treat them according to their relevance and to see which the best are. For this, we use the objective function. This function measures the performance of each individual. To be able to judge the quality of an individual and thus compare him to others. The objective function of our case is to minimize the cost while maximizing reliability. The evaluation of each individual in the population then makes it possible to make the selection. For a system made up of n components in parallel, the reliability to be maximized is given by equation 2. (2) ? ? ? ? ? ? ? ? j j j j j l u l a a a a ? ? ? ? ? ? ? ? ? , jj lu aa 1 ,..., ,..., i j m X a a a ?? ? ?? ? ? 1 11 n i i Rr ? ? ? ? ? Where: n is the number of components r i is the reliability of the component i For a parallel-series system, the reliability to be maximized is given by equation 3. (3) Where: p is the number of stages of the system r ij the reliability of the j th element of the i th stage n i is the number of components of the i th stage The cost function to be minimized deducted from the work of (Gutha and Vadlamani [9]) for a parallel -series system (machine) with p stages is gave by equation 4. (4) Where: -C i is the cost vector of the components of the X i chromosome p is the number of stages of the machine. # iv. Selection of the most suitable individuals When the entire population is assessed at generation t, individuals are ranked in ascending order of objective function. Then the selection is made. Selection helps to statistically identify the best individuals in a population and eliminate the bad ones from one generation to the next. This operator also gives a chance to the bad elements because these elements can, by crossing or mutation, generate relevant descendants compared to the optimization criterion. The first N × G individuals (the best N × G ) are selected to be parents. G is the third setting parameter of the genetic method. G is called the generation gap. G makes it possible to select a part of the population to provide sufficient genetic material without decreasing the speed of convergence (Goldberg [6]). # There are different selection techniques: ? Selection by rank: This selection method always chooses the individuals with the best adaptation scores, without allowing chance to intervene; ? Selection by wheel: For each parent, the probability of being selected is proportional to their adaptation to the problem (their score by the fitness function). This selection be imaged by a casino roulette wheel, on which all the chromosomes of the population are placed, the place is given to each of the chromosomes being proportional to its adaptation value. Also, the higher an individual's score, the more likely he is to be selected. We spin the wheel as many times as we want individual sons. The best will be able to be drawn several times, and the worst never; ? Selection by tournament: Two individuals are chosen at random, their adaptation functions are compared, and the best suited is selected.; # C C X e such that each individual has the same probability P(i) = 1 / N as all other individuals, where N is the total number of individuals in the population; ? Elitism: The passage from one generation to another through the crossing and mutation operators creates a great risk of losing the best chromosomes. Therefore, elitism aims to copy the best (or first -best) chromosome (s) from the current population to the new population before proceeding to the mechanisms of crossing and mutation. This technique quickly improves the solution because it prevents the loss of the most qualified chromosome when passing from one generation to another. # v. Crossing The selected population is divided into N / 2 couples formed randomly. Two parents, and are chosen randomly from the potential parents P 1 and P 1 their genes are combined according to equation 5. (5) Where: ? j is a uniform random number, -k = N × G +1,..., N, the k th individual, mode can j = 1,..., m. The newly created individual is then evaluated. If its adaptation is better than that of the worst parent, it is integrated into the population to training the next generation. If it is not the case, we repeat the combination. # vi. Mutation of all individuals in the population The mutation operator is a process where a minor change in the genetic code is applied to an individual to introduce diversity and thus avoid falling into local optima. This operator is applied with a probability P m . P m generally lower than that of the crossing P c . This probability must be low. Otherwise the GA will turn into a random search. algorithm, the parameters N and G must be adjusted with care. The size of the population N affects both the performance and efficiency of the algorithm (Bicking et al. [10]). The algorithm is less efficient with very small population sizes. Large population size may contain more interesting solutions and discourage premature convergence towards sub-optimal solutions, but requires more assessments per generation, which can lead to a low convergence rate. The generation gap G determines the proportion of the population that remains unchanged between two generations. It is chosen to select individuals as severely as possible, without destroying the diversity of the population too much. The global strategy used assumes that all the individuals who make up the population, from generation to generation, satisfy all the constraints. The best solution for the latest generation represents the solution to the problem by the defined criteria. # III. # Application Consider a machine made up of five components mounted in parallel (see Figure 1). Our goal is to define an optimal reliability network, to ensure operation with minimum reliability r min = 0.80 while minimizing the cost of the structure. Step 1: Data coding We choose binary coding using a 5-bit character string. Bits 1 through 5 represent the components of C 1 through C 5 , respectively. If a component exists in the generated solution, then its ? ? ? ? ? ? ? ? 1 21 () j # vii. Choosing the best solutions This choice consists in retaining the solutions which have a lower value of the objective function, and putting them in the population P k+1 . # viii. Stopping criterion The stopping criterion is evaluated in the current population. If it is filled, the whole population has converged on the solution. Otherwise the reproduction pattern will be repeated. The stopping criterion used in this method expresses that all individuals have converged on the same solution and assumes that evolution is no longer possible, that is to say, that no better solution can be found. The whole strategy is elitist because only the best individuals are selected for survival from one generation to the next and can be the parents of new and better individuals. To ensure convergence of the corresponding bit takes the value "1"; otherwise, it takes the value "0". Step 2: Choice of GA parameters This choice is random while waiting to improve it afterward. The parameters of the GA are as follows: -The size of the population: N = 4 ; -The number of generations: K = 10 ; -The probability of crossing: P c = 0.6 ; -The probability of mutation: P m = 0.01. Step 3: Generation of the initial population We randomly generate a population of 4 noted chromosomes X 1 , X 2 , X 3 and X 4 : Step 4: Definition of objective function The objective function f to optimize is defined by equation 6. (6) With: X 1 = [01001], X 2 = [10110], X 3 = [10011] and X 4 = [11001]. r min : the minimum reliability of the structure to be optimized; Step 5: Evaluation of the reliability of each X i chromosome in the population For each X i chromosome generated, the reliability and the cost are evaluated respectively by relations 2 and 4. So: ---- Step 6: Selection of the most suitable chromosomes The best chromosomes in terms of cost in descending order are: X 1 , X 2 , X 3 and X 4 . We select the chromosomes X 1 , X 2 and X 3 for reproduction and we eliminate the chromosome X 4 because it is the worst as far as cost is considered. Step 7: Crossing We cross a couple among the selected chromosomes with a crossover rate of 0.6 to form new children. Consider the couple (X 1 , X 2 ): Step 8: Mutation We do a random draw of a single chromosome gene X' 1 , X' 2 and X 3 ; then the selected bit will be mutated with a mutation probability of 0.01. Consider the second bit of the chromosome X' 2 : Step 9: Substitution We are replacing the new population with the new chromosomes. Chromosomes X 3 and X 4 will is eliminated because they are the least suitable, and the chromosome X 1 and X 2 will be kept .After all, they are the best qualified among the individuals of the population. So the new population will be made up of chromosomes X 1 , X 2 , X' 1 and X'' 2 . X 1 = [01001], X 2 = [10110], X' 1 = [01010] and X'' 2 = [11101]. Step 10: Repeat steps 4 to 9 The algorithm stops either after the reproduction of 16 generations or when we notice that the solution does not improve after a defined number of generations. Step 11: Optimal solutions -The corresponding chromosome that optimize objective function is: -The optimum reliability and costs for each solution are: By considering the components selected by the genetic algorithm in the previous optimal solutions, we generate all the possible paths (connections) between the component of the machine. In solution 1 (X a chromosome) , we retain the components in parallel C 1 and C 5 ; in solution 2 , we retain the components C 1 and C 2 (thus defining a path); in solution 3, we retain the components C 1 and C 3 (defining a path) and from end to end until the last solution ( X g chromosome), the last path connecting the components C 3 and C 4 is shown. From where the network of optimal reliability of the machine is given by Figure 3. By evaluating the overall reliability of the reliability network, we obtain a reliability of 0.88, higher than the minimum reliability set. Similarly, by evaluating the overall cost of the machine by applying relation 4, for the reliability network we obtain a cost of 88.75, which is much lower than the overall cost generated by the initial system, which is 103.75. Objectively, the reliability network has favored cost reduction on the one hand reliability increase on the other. ? ? ? ð??" F E ? ? ii i min minimize f(X ) = C(X ) s.t r X r ? ? ?10 0 10 XX X X ? ?? ? ?? ? ?? ? ? ? ? ? ? ? ? ? ? ' '' 22 1 101 1 0 01 11 XX ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? 10001 , 11000 , 10100 , 01001 , 00101 , 01010 a b c d e f X X X X X X ? ? ? ? ? ? ? ? 0. # IV. # Conclusion Having completed the writing of this paper, which concerns the optimization of the reliability of a system by the genetic algorithm, it appears that the general objective has been achieved. Indeed, through the functionality of genetic algorithms, implemented on a structure with five components, we were able to reduce the cost and increase the reliability of the parallel structure. However, although the genetic algorithm is easy to implement, it might require an infinite number of iterations to the best approach of optimal solution. This would make the algorithm less robust. In perspective for this work, it would be wise to develop or implement another meta-heuristic that would not require a large number of iterations. # Data Availability The data needed to support these results are available in the text, specifically in Table 1. Conflicts of Interest All authors declare that there are no conflicts of interest regarding the publication of this paper. 1 1ComponentsReliabilityCost????0.9515????0.9717????0.914????0.8912??ð??"ð??"0.9616 © 2020 Global Journals Machine Reliability Optimization by Genetic Algorithm Approach ## Acknowledgments This work was funded by the authors of the paper and was carried out in the Department of Fundamental Sciences and Engineering Techniques at the Chemical Engineering and Mineral Industries School at the University of Ngaoundere. * Genetic algorithm in optimization of system reliability LPainton JCampbell IEEE Trans Reliab 44 1995 * Optimizing survivability of vulnerable series-parallel multi-state systems GLevitin ALisnianski Reliability Engineering and System Safety 79 2003 * Cybernetic solution path of an experimental poblem'', Library Translation 1122 IReichenberg Royal Aircraft Establishment 1965 * Adaptation in natural and artificial systems JHHolland 1975 University of Michigan press * Genetic algorithms DGoldberg 0-201-15767-5, 1989a Addison Wesley, ISBN * Genetic algorithms in search, optimization and machine learning DGoldberg 1989b Addison Wesley * Course of Heuristic and Metaheuristic Exact Resolution Methods SMDouiri SElbarnoussi HLakhbab 2013 Mohammed V University, Faculty of Sciences of Rabat, Morocco, Mathematics, Computer Science and Applications Research Laboratory * Numerical Optimization of computer models HPSchewefel 1981 Editions Wiley * Modified Harmony Search Applied to Reliability Optimization of Complex Systems JGutha RVadlamani Advances in Intelligent Systems and Computing 382 2016 * Global optimization by artificial life: a new technique using genetic population evolution FBicking CFonteix JPCorriou IMarc RAIRO-Operations Research 28 1 1994