Teshale Shinato
Transcription
Teshale Shinato
SOME ITERATIVE METHODS FOR SOLVING SYSTEM OF NONLINEAR EQUATIONS MSc PROJECT TESHALE SHINATO MAY 2016 HARAMAYA UNIVERSITY, HARAMAYA Some Iterative Methods for Solving System of Nonlinear Equations A Project Submitted to the Department of Mathematics, Postgraduate Program Directorate HARAMAYA UNIVERSITY In Partial Fulfillment of the Requirements for the Degree of MASTER OF SCIENCE IN MATHEMATICS (NUMERICAL ANALYSIS) Teshale Shinato May 2016 Haramaya University, Haramaya HARAMAYA UNIVERSITY Postgraduate Program Directorate I hereby certify that I have read and evaluated this Project titled ‘Some Iterative Methods for Solving System of Nonlinear Equations’ prepared under my guidance by Teshale Shinato. I recommend that it be submitted as fulfilling the project requirement. Seleshi Demie (PhD) Advisor __________________ Signature ___________________ Date As member of the board of Examiners of the M.Sc Project Open Defense Examination, we certify that we have read and evaluated the Project prepared by Teshale Shinato and examined the candidate. We recommend that the Project be accepted as fulfilling the Project requirement for the degree of Master of Science in Mathematics (Numerical Analysis). Chairperson ________________ Signature ________________ Date ________________ Internal Examiner ________________ Signature ________________ Date ________________ External Examiner ________________ Signature ________________ Date Final approval and acceptance of the project is contingent upon the submission of its final copy to the Council of Graduate Studies (CGS) through the candidate’s department or school of graduate committee (DGC or SGC). ii STATEMENT OF THE AUTHOR By my signature below, I declare that this Project is my own work. I have followed all ethical and technical principles of scholarship in the preparation, and compilation of this Project. Any scholarly matter that is included in the Project has been given recognition through citation. This Project is submitted in partial fulfillment of the requirements for the degree of Master of Science in Mathematics (Numerical Analysis) at the Haramaya University. The Project is deposited in the Haramaya University Library and is made available to borrowers under the rules of the Library. I solemnly declare that this Project has not been submitted to any other institution anywhere for the award of any academic degree, diploma or certificate. Brief quotations from this Project may be made without special permission provided that accurate and complete acknowledgement of the source is made. Requests for permission for extended quotations from or reproduction of this Project in whole or in part may be granted by the Head of the Department when in his or her judgment the proper use of the material is in the interest of scholarship. In all other instances, however, permission must be obtained from the author of the Project. Name Date Department: Mathematics Signature iii ABBREVIATIONS BM Broyden’s method NI Number of Iterations NM Newton’s method PCM-1 The predictor corrector method-1 PCM-2 The predictor corrector method-2 SNLEs System of nonlinear equations iv BIOGRAPHICAL SKETCH The author was born in 1989 on April 16 in SNNP Regional State, Hadiya Zone, Gibe wereda, Megacho kebele. He attended his primary education at Muma primary school and joined Yekatit 25/67 Secondary School for grades 9 and 10, and Wachamo Comprehensive Secondary and Preparatory School for grades 11 and 12 which are found at the Hosanna town. He then joined Dire Dawa University in 2007 and received Bachelor of Education degree in Mathematics in July, 2009. He has been employed in Southern Regional State, Hadiya Zone as a High school teacher on August 15, 2009 and stayed at work for four years in the school. After working for four years in that school, he then joined Postgraduate Program at Haramaya University, College of Natural and Computational Sciences, Department of Mathematics on October, 2013 to pursue a program of study for MSc degree in Mathematics (Numerical Analysis). v ACKNOWLEDGEMENT I would like to express my sincere gratitude to my advisor Dr. Seleshi Demie next to Almighty God for his profound commitment and encouragements he has shown throughout my study period. I pay respect and express indebtedness to him because of his inspiring guidance, consistent supervision especially his suggestions in selection of tools from the grass root of the proposal and at every phase of this project work. Without his knowledge, direction, guidance, and all of his help, this project would not have been achieveable. My gratitude goes to Gibe Woreda, particularly, to the Woreda’s Education Sector for the chance they have given me without any opposition and accepting my questions for joining the postgraduate program to study M.Sc. in mathematics and for their provision of the necessary support to complete the postgraduate study at Haramaya University. I would also like to acknowledge how thankful I am for my Mom and Dad. Without their unconditional love and support, and always being by my side with prayer and encouragement throughout the years, it would have been so much difficult to be successful. Also I have a heartfelt gratitude for all my brothers and sisters especially for the two elder brothers, Tade and Fike for their support in all necessary conditions. I strongly understood that it was impossible to achieve a lot of my goals without their support. vi TABLE OF CONTENTS STATEMENT OF THE AUTHOR iii ABBREVIATIONS iv BIOGRAPHICAL SKETCH v ACKNOWLEDGEMENT vi LIST OF TABLES ix ABSTRACT x 1.INTRODUCTION 1 1.1.Background of the Study 1 1.2.Statement of the Problem 3 1.3.Objectives of the Study 4 2.LITERATURE REVIEW 6 2.1.Iterative Methods for Solving SNLEs 6 2.2.Quadrature Formula Based Methods for Solving SNLEs 8 3.MATERIALS AND METHODS 10 4.SOME PRELIMINARY CONCEPTS 11 4.1.Some Relevant Terms 11 4.2.Some Quadrature Formulas 16 5.SINGLE STEP ITERATIVE METHODS FOR SOLVING SNLEs 17 5.1.Newton Method 17 5.2.Broyden’s Method 21 6.TWO ITERATIVE METHODS WITH TWO STEPS FOR SOLVING SNLEs 27 6.1.Predictor-corrector Methods 1 and 2 27 6.2.Convergence Analysis of the Two Methods 30 Continues… vii 6.2.1.Proof of cubic order of convergence for algorithm 6.2 30 6.2.2.Proof of cubic order of convergence for algorithm 6.4 36 7.NUMERICAL RESULTS OF THE FOUR METHODS 45 8.SUMMARY, CONCLUSION AND RECOMMENDATION 61 8.1.Summary 61 8.2.Conclusion 62 8.3.Recommendation 62 9.REFERENCES 63 10.APPENDICES 66 viii LIST OF TABLES Table Page 1: Iteration result of Newton’s method for Example 7.1 46 2: The iteration result of PCM-1 for Example 7.1 49 3: The iteration result of PCM-2 for Example 7.1 51 4: The summary of the iteration results for Example 7.1 by Broyden’s method 53 5: Comparison of BM, NM, PCM-1 and PCM-2 for Example 7.1 53 6: The summary of iteration results of Newton’s method for Example 7.2 54 7: The summary of iteration results for PCM-1 in Example 7.2 55 8: The iteration result of PCM-2 for the system given in Example 7.2 55 9: The iteration result of the Broyden’s method for the system given in Example 7.2 55 10: Comparison of BM, NM, PCM-1 and PCM-2 for Example 7.2 56 11: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.3 57 12: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.4 58 13: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.5 59 14: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.6 60 ix Some Iterative Methods for Solving System of Nonlinear Equations ABSTRACT System of nonlinear equations arises in different areas of applied sciences such as engineering, physics, chemistry, medicine, and robotics. They appear also in many geometric computations such as intersections, minimum distance, and when solving initial or boundary value problems in ordinary or partial differential equations. Solving system of nonlinear equations is difficult when compared to solving its linear counterpart and finding solutions of nonlinear problems is important in the study of numerical analysis. Iterative methods such as Newton’s method, Broyden’s method, Adomian decomposition method, Steepest Descent method, Homotopy perturbation method, predictor-corrector methods, etc. are among the methods used to solve nonlinear equations. This project investigated solutions of SNLEs by using Newton’s and Broyden’s methods which are single step methods and two methods with two steps each which are predictor-corrector method-1 and 2 (PCM-1 and 2). For both PCM-1 and PCM-2, Newton’s method served as a predictor and algorithms 6.3 and 6.5 served as corrector, respectively. The comparisons made among these methods were based on the number of iterations they took to converge to the solution by considering numerical examples from SNLEs. The results indicated that PCM-1 and PCM-2 took less number of iterations to converge and hence they performed better than Newton and Broyden’s methods. x 1. INTRODUCTION 1.1. Background of the Study Numerical analysis is the area of mathematics that creates, analyzes, and implements algorithms for solving numerical problems. Such problems originate generally from real-world applications of algebra, geometry and calculus. These problems also occur throughout the natural sciences, social sciences, engineering, medicine, and business. During the past half-century, the growth in power and availability of digital computers has led to an increasing use of realistic mathematical models in science and engineering, and highly sophisticated methods of numerical analysis have been needed to solve these more detailed mathematical models of the world. The formal academic area of numerical analysis varies from quite theoretical mathematical studies (Atkinson and Han, 2001) to computer science issues (Edwards, 1997). With the growth in importance of using computers to carry out numerical procedures in solving mathematical models of the world, an area known as scientific computing or computational science has taken shape during the 1980s and 1990s. This area looks at the use of numerical analysis (Fosdick et al., 1996) from a computer science perspective. It is concerned with using the most powerful tools of numerical analysis, computer graphics, symbolic mathematical computations, and graphical user interfaces to make it easier for a user to set up, solve, and interpret complicated mathematical models of the real world. Therefore, numerical analysis is one of the fields of mathematics which uses different types of methods to solve complicated mathematical problems in the areas stated above. There are different kinds of challenging problems in science, engineering and other fields of study which need to be solved. The most important problem among those problems is finding the solution for system of nonlinear equations. There are simple formulas for solving linear and quadratic equations; and there are also somewhat complicated formulas for cubic equations. Besides polynomial equations, we have many problems in scientific and engineering applications that involve the functions of transcendental nature. 2 Numerical methods are often used to obtain approximated solutions of some complicated problems because those problems are not possible to obtain exact solution by usual algebraic processes. The solutions of systems of equations have a well-developed mathematical and computational theory when solving linear systems, or a single nonlinear equation. The situation becomes complicated when the equations in the system do not exhibit nice linear or polynomial properties. In this case, both the mathematical theory and computational practices are far from complete understanding of the solution process. So systems of nonlinear equations are one of such types of problems. They arise in various domains of practical importance such as engineering, medicines, chemistry, and robotics. They appear also in many geometric computations such as intersections, minimum distance, and when solving initial or boundary value problems in ordinary or partial differential equations. As it is mentioned above, solving systems of nonlinear equations is more difficult when compared with the linear counterpart. So finding solutions of these problems is one of the most important parts of the studies in numerical analysis and because of this reason different mathematicians developed different iterative methods to solve problems of systems of nonlinear equations. For example, there are the following types of methods: methods derived from Quadrature formulas, Newton’s method, Broyden’s method, Adomian decomposition method, Steepest Descent method, Homotopy perturbation method, predictor-corrector methods, etc. This project paper mainly focused on Newton’s method, Broyden’s method and two iterative methods with two steps each (predictor-corrector methods 1 and 2) for solving system of nonlinear equations. Historically Newton’s method was developed by Sir Isaac Newton around 1600. Most experts agree that the road for the Newton’s method was opened by Francois Vieta in 1600 (Deuflhard, 2012). Isaac Newton was using this method to approximate numerical solutions of scalar nonlinear equations. One of his first applications of this was on the numerical solution of a cubic polynomial equation 𝑓(𝑥) = 𝑥 3 − 2𝑥 − 5 = 0 at 𝑥 (0) = 2 (Deuflhard, 2012). This 3 method later developed by Thomas Simpson after the derivatives concepts introduced. Simpson played a great role for leaping Newton’s method from using on scalar equations to system of equations (Deuflhard, 2012). Broyden’s method is an iterative method developed by Charles Broyden around 1665. The reason for the development of this method was the drawbacks in Newton’s method. Broyden thought (Deuflhard, 2012) that the evaluation of Jacobian matrix at every iteration is a difficult task and he need to overcome this problem. Then Broyden started the iteration by computing the whole Jacobian only at the first iteration and then to use the rank-one updated approximate matrix at the other iterations. The Predictor-corrector methods 1 and 2 are iterative methods which are derived from closedopen quadrature formulas by using Taylor polynomial. It is well known that the quadrature rules play an important and significant role in the evaluation of the integrals (Burden and Faires, 2001; Noor, 2007a, b; Podisuk et al., 2007; Darvishi and Barati, 2004; Frontini and Sormani, 2004; Cordero and Torregrosa, 2006; Cordero and Torregrosa, 2007; Darvishi and Barati, 2007; Babajee et al., 2008). It has been shown that the quadrature formulas can be used to develop some iterative methods for solving a system of nonlinear equations. The two iterative methods with two steps each in this project paper are also methods developed from quadrature formulas. 1.2. Statement of the Problem System of nonlinear equations is a problem that arises in different areas of studies as mentioned in section 1.1. Finding solutions for such problems are not easy when compared with other problems such as solving single nonlinear equations, linear equations, quadratic equations, polynomial equations, equations of transcendental nature, system of linear equations, and so on. Because of this a number of mathematicians tried to find different iterative methods to solve problems of system of nonlinear equations. From those methods, this project focused on four iterative methods from which two of them are methods with a single step (i.e., Newton and Broyden’s methods) and the remaining two methods are methods 4 with two-steps each (i.e., predictor-corrector methods 1 and 2). For both predictor-corrector methods 1 and 2, the predictor step is a Newton’s method and the corrector steps are the two iterative methods, which were derived from quadrature formulas using Taylor series as it has been shown in algorithms 6.3 and 6.5 of chapter-6. So these are methods to solve SNLEs of the form: 𝑓1 (𝑥1 , 𝑥2 , 𝑓2 (𝑥1 , 𝑥2 , 𝑓3 (𝑥1 , 𝑥3 , ⋯ , 𝑥𝑛 ) = 0 𝑥3 , ⋯ , 𝑥𝑛 ) = 0 𝑥3 , ⋯ , 𝑥𝑛 ) = 0 𝑥2 , ⋮ 𝑓𝑛 (𝑥1 , 𝑥2 , 𝑥3 , ⋯ , 𝑥𝑛 ) = 0 where each function 𝑓𝑖 can be thought of as mapping a vector X = [𝑥1 , 𝑥2 , 𝑥3 , ⋯ 𝑥𝑛 ]𝑇 of the 𝑛 dimensional space Rn into the real line R, for 𝑖 = 1,2,3, … . The general system of nonlinear equations in 𝒏 unknowns can be alternatively represented by a function F, mapping from R n into Rn, as 𝑓1 (𝑥1 , 𝑥2 , 𝑥3 , … … , 𝑥𝑛 ) 0 𝑓2 (𝑥1 , 𝑥2 , 𝑥3 , … … , 𝑥𝑛 ) 0 F(𝑥1 , 𝑥2 , 𝑥3 , … … , 𝑥𝑛 ) = 𝑓3 (𝑥1 , 𝑥2 , 𝑥3 , … … , 𝑥𝑛 ) = 0 ⋮ ⋮ [ 0] [𝑓𝑛 (𝑥1 , 𝑥2 , 𝑥3 , … … , 𝑥𝑛 )] In a vector notation system, equation (2) can be written as: F(X) = 0 The functions, f1(x1, x2, …, xn), f2(x1, x2, ..., xn), f3(x1, x2, . . . ., xn), . . .,and fn(x1, x2, …, xn) are called the coordinate functions of F(X). The problem of solving a nonlinear system of equations is equivalent to the statement: For F: D⊆Rn → Rn , find x ∗ ∈ Rn such that F (x ∗ )= 0, where F(x) is some nonlinear vectorvalued function with n unknowns. 1.3. Objectives of the Study The general objective of this project is to study four iterative methods (i.e., Newton’s method, Broyden’s method, predictor-corrector method-1 and predictor-corrector method-2), which are 5 used to solve SNLEs and to compare the performance of these methods one with the other based on the number of iterations taken to reach to the approximate solution. The study is intended to achieve the following specific objectives; To derive the iterative schemes for the four methods. To prove the cubic order of convergence for the four methods. To use these methods to solve system of nonlinear equations. To compare the four methods based on the number of iterations they took to converge to the solution. 6 2. LITERATURE REVIEW 2.1. Iterative Methods for Solving SNLEs In order to solve the problems of SNLEs encountering in everyday life activities, many numerical iterative methods were developed by different researchers. In this section we deal with some of these methods. Darvishi (2009) used a two-step higher order Newton-like method for solving systems of nonlinear equations. The method considers the problem of finding a real root of a function F: ̅ of the system of nonlinear equations F(x) = 0 of n equations Rn →Rn, that is a real solution, 𝒙 with n variables. This solution can be obtained as a fixed point of some function G: Rn →Rn by means of the fixed point iteration method. Darvishi and Barati (2007) derived a super cubic iterative method from the Adomian decomposition method to solve systems of nonlinear equations. The authors showed that the method is third-order convergent using classical Taylor expansion but Babajee et al. (2008) introduced a note on the local convergence of iterative methods based on Adomian decomposition method and 3-node quadrature rule. They formulated their method by using Ostrowski’s technique based on point of attraction to solve system of nonlinear equations. The authors showed that the method of Darvishi and Barati (super cubic iterative methods to solve systems of nonlinear equations) was in fact the fourth order convergent not the third order. They pointed out that the new method is better than that of the classic Darvishi and Barati methods after the proof of point of attraction theorem. Noor et al. (2015) introduced a new family of iterative methods to solve a system of nonlinear equations, F(X) = 0 by using a new decomposition technique which is quite different from Adomian decomposition method. In the implementation of Adomian decomposition, one has to calculate the derivatives of the so-called Adomian polynomial, which is itself a difficult problem. To overcome the difficulty of Adomian method they used the decomposition of Daftardar-Gejji and Jafari (2006) as developed by Noor et al. (2006) to suggest new iterative methods for solving system of nonlinear equations. Their method does not involve the higher- 7 order differentials of the functions and is very simple as compared with the Adomian decomposition. Furthermore, by improving Newton’s method, Chun (2006) has presented a new iterative method to solve nonlinear equations. By using Adomain decomposition, Darvishi et al. (2007a; b) have constructed new methods, and Golbabai and Javidi (2007) have applied the homotopy perturbation method to build a new family of Newton-like iterative methods for solving system of nonlinear equations. Ozel (2010) has considered a new decomposition method for solving the system of nonlinear equations. Hafiz and Bahgat (2012) modified Householder iterative method for solving system of nonlinear equations. Broyden, C. (1965) formulated a very important iterative method, the so called a class of iterative methods for solving system of nonlinear equations. The reason that made him to introduce this method is because of the challenges that faced him when solving problems of system of nonlinear equations by using Newton’s method. He was well aware of the shortcomings of Newton’s method and thought how to overcome the problem, especially, the problem in computing the Jacobian matrix and finding its inverse at each and every iteration which was difficult, time consuming and expensive operation, particularly, during solving large sized system of nonlinear equations. This method has preferred to use an updating approximation matrix instead of the Jacobian matrix. By doing this, Broyden simplified the cost of computation of Jacobian matrix in the iteration process of solving SNLEs. There are advantages and some drawbacks of this method. The advantage of Broyden’s method is the reduction of computations, specifically the way that the inverse of the approximation matrix can be computed directly from the previous iteration for the next step reduces the number of computations in comparison to Newton’s method. The drawback of this method is that the method does not converge quadratically. This means that more iteration number may be needed to reach the solution, when compared to the number of iterations that the Newton method requires. Anuradha and Jaiswal, (2013) formulated the third and fourth order iterative methods for solving systems of nonlinear equations. The main objective of their work was to extend the third order iterative method for solving single nonlinear equations into system of nonlinear equations. They were also motivated to develop the method which improves the order of 8 convergence of Newton’s method with minimum number of functional evaluations. To achieve this goal, Anuradha and Jaiswal used a weighted method. To compare their computational efficiency they used not only traditional ways but also recently introduced flops-like based concepts. Finally, from their work they found that the efficiency of quadratically convergent multidimensional method is not satisfactory in the most of the practical problems so that the method extended from the third order method for solving single nonlinear equations into multivariate case to construct the fourth order convergence by using weighted method concept and it is much more better than the other compared methods. The sense of their efficiency index was not dependent on only the number of functional evaluations but also on the number of operations per iterations. They have given the comparison of efficiencies based on flops and functional evaluations. They showed that the traditional efficiency index of the fourth order method is same but flops-like efficiency index of their method dominates the other methods. Hosseini (2010) developed an improved homotopy analysis method for solving a system of nonlinear algebraic equations and compared this method with Newton-Raphson method and homotopy perturbation method. They have shown that this proposed method is very effective and convenient in performance than the other two methods, namely Newton-Raphson and homotopy perturbation. The proposed method converges fast to the solution as compared with Newton-Raphson and homotopy perturbation methods. 2.2. Quadrature Formula Based Methods for Solving SNLEs Mamat et al. (2014) formulated a trapezoidal Broyden’s method for solving SNLEs. In this method, the authors used the trapezoidal rule to solve the problems and in the method, the two point predictor-corrector approach was used, where the Broyden’s method is the predictor and the proposed method is the corrector. They compared the performance of the proposed method with three Newton-like methods namely Newton’s method (NM), Broyden’s method (BM) and Fixed-Newton’s method (FM) and showed the performance of the proposed method to be better than the methods compared with it because of taking fewer numbers of iterations until the process converges. Hafiz and Bahgat (2012) proposed and analyzed a new method the so called an efficient two step-iterative method for solving system of nonlinear equations. This is 9 a predictor-corrector method developed by using the weight combination mid-point, trapezoidal and quadrature formulas. They used a Newton’s method as a predictor step and the new method as the corrector step of their method. Hafiz and Bahgat compared their method with a Homotopy perturbation and Newton-Raphson methods based on their performance. Some numerical examples were used to show the efficiency of the method relative to the compared methods. The result of the comparison indicates that the method developed by Hafiz and Mohamed is very effective and provided highly accurate results in a less number of iterations than that of Homotopy perturbation’s method and Newton-Raphson method. Frontini and Sormani, (2004) developed third-order methods from quadrature formulae for solving SNLEs. Their method is the extension from single dimensional to n-dimensional problems of system of nonlinear equations. They made a modification on a Newton’s method based on quadrature formulas of order at least one, which produces iterative methods with order of convergence three. In this method the authors have given a general error analysis providing the higher order of convergence. The effectiveness of their methods is tested by taking some counter examples from system of nonlinear equations. The comparison is also method with some other existing methods to show how fast and effective are the methods. Hassan and Waziri (2015) presented a Broyden-like update via some quadrature formulas for solving system of nonlinear equations. Their method is a new alternative approximation based on the quasi-Newton approach for solving SNLEs using the average of midpoint and Simpson’s quadrature formulas. The goal of this method is to enhance the efficiency of Broyden’s method by reducing the number of iterations it takes to reach the solution. The authors have given the local convergence analysis and computational results showing the relative efficiency of the proposed method. The computational experiment of the method is given by using some numerical examples and compared to the existing classical methods. Finally, the overall results of this new method imply that the proposed method is effective and better than that of original Broyden’s method. In this project, four iterative methods (Newton’s method, Broyden’s method, predictorcorrector methods 1 and 2) to solve a system of nonlinear equations were discussed and finally some numerical examples were used to demonstrate the performance of the methods and compared them depending up on the number of iterations. 10 3. MATERIALS AND METHODS This project work has been developed through a number of reference books and materials which are available in the library of Haramaya and Addis Ababa universities and also collected some quite relevant information from the soft copies, internet, etc. Related journals, seminars and projects were examined in detail and used to consolidate all the entire frameworks and skeleton of the project. Some necessary definitions, theorems, and examples are contained in this project to illustrate the essential idea of iterative methods for solving system of nonlinear equations. MATLAB programs were used to minimize the high consumption of time and energy that could happen during computations by hands and to accomplish the work easily. 11 4. SOME PRELIMINARY CONCEPTS In this chapter, some preliminary concepts, relevant definitions and well known results, such as norm, convex set, Frèchet derivative, lipschitz continuity, convergence of iterative methods, stopping criteria, and some quadrature formulas have been presented to be helpful throughout the study of this project. 4.1. Some Relevant Terms Iterative methods: The term iterative method refers to a procedure that is repeated over and over again, to find the root of an equation or find the solution of a system of equations. It is a technique that uses successive approximations to obtain more accurate solutions to systems of equations at each step. In a computational mathematics, an iterative method is a mathematical procedure that generates a sequence of improving approximate solutions for a class of problems. A specific implementation strategy of an iterative method, including the termination criteria, is an algorithm of the iterative method. Algorithms: In mathematics and computer science, an algorithm is a step-by-step procedure for calculations. Algorithms are used for calculation, data processing, and automated reasoning. More precisely, an algorithm is an effective method expressed as a finite list of well-defined instructions for calculating a function starting from an initial state and initial input perhaps empty. The instructions describe a computation that, when executed, will proceed through a finite number of well-defined successive states, eventually producing "output" and terminating at a final ending state. 12 Vector Norm: A vector norm on Rn is a function,‖. ‖ from Rn into R with the following properties: (i) ‖X‖ ≥ 0 for all X∈Rn (ii) ‖X‖ = 0 if and only if X=0 (iii) ‖𝛼X‖ = |𝛼|‖X‖ for all 𝛼 ∈R and X∈Rn ‖X + 𝑌‖ ≤ ‖𝑋‖ + ‖𝑌‖ for all X, 𝒀 ∈ Rn (iv) Vectors in Rn are column vectors, and it is convenient to use the transpose notation when a vector is represented in terms of its components. For example, the vector 𝑥1 𝑥2 X= 𝑥3 , can be written as X= (𝑥1 , 𝑥2 , 𝑥3 , … , 𝑥𝑛 )T. ⋮ [𝑥𝑛 ] If X= (𝑥1 , 𝑥2 , 𝑥3 , … , 𝑥𝑛 )T ∈Rn is a vector, then 1 (a) 𝑙𝑝 -norm of X is, ‖X‖p = {∑𝑛𝑖=1|𝑥𝑖 |𝑝 }𝑝 , 1≤ 𝑝 < ∞ If p = 1, then ‖X‖1 = |𝑥1 | + |𝑥2 | + |𝑥3 | + ⋯ + |𝑥𝑛 |. (b) The 𝑙2 −norm for the vector X is called the Euclidean norm because it represents the length of the vector denoted by ‖X‖ = ‖X‖2 = √|𝑥1 |2 + |𝑥2 |2 + |𝑥3 |2 + ⋯ + |𝑥𝑛 |2 (c) 𝑙∞ - Norm of X represents the absolute value of the largest component in the vector X. It is denoted by, ‖X‖∞ = max |𝑥𝑖 | 1≤𝑖≤𝑛 Matrix Norm: A matrix norm on the set of all n×n matrices is a real valued function, ‖. ‖ defined on this set, satisfying the following properties for all n×n matrices A and B and a real number 𝛽: (i) ‖A‖ ≥ 0 (ii) ‖A‖ = 0, If and only if A is 0, the matrix with all entries equal to zero. (iii) ‖𝛽A‖ = |𝛽|‖A‖ for all 𝛽 ∈ R, (iv) ‖A + B‖ ≤ ‖A‖ + ‖B‖, 13 (v) ‖AB‖ ≤ ‖A‖‖B‖. The distance between 𝑛 × 𝑛 matrices A and B with respect to this matrix norm can be given as‖A − B‖. If A = [𝑎𝑖𝑗 ] is an n×n matrix, then one can define the norm of matrix A as: (a) 𝑙1− norm ‖A‖1 = max ∑n𝑖=1|𝑎𝑖𝑗 | 1≤𝑗≤3 (b) 𝑙∞ − norm ‖A‖∞ = max ∑𝑛𝑗=1|𝑎𝑖𝑗 | 1≤𝑖≤𝑛 (c) 𝑙2 − norm 2 ‖A‖2 = √∑𝑛𝑖=1 ∑𝑛𝑗=1(𝑎𝑖𝑗 ) = √∑𝑛𝑖=1{(𝑎𝑖1 )2 + (𝑎𝑖2 )2 + (𝑎𝑖3 )2 + ⋯ + (𝑎𝑖𝑛 )2 } Sometimes there are functions that can be written in a matrix notation form when the functions are as in the case of system of equations. So system of nonlinear equations having of the form F(x1,x2 , x3 ,. . .,xn)= [f1(x1, x2, …, xn), f2(x1, x2, …, xn), . . .,fn(x1, x2, …, xn) ]T can be one of such equations and written in a matrix form as: f1 (x1 , x2 , x3 , … … , xn ) x1 f2 (x1 , x2 , x3 , … … , xn ) x2 (x ) f , x , x , … … , x F(X) = 3 1 2 3 n , where X= x3 ⋮ ⋮ [xn ] [fn (x1 , x2 , x3 , … … , xn )] Jacobian Matrix: A jacobian matrix is an 𝑛x𝑛 matrix of first order partial derivatives of component functions of a system of nonlinear equations that can be written as 𝜕𝑓1 𝜕𝑓1 𝜕𝑓1 𝜕𝑥1 𝜕𝑓2 𝜕𝑥2 𝜕𝑓2 𝜕𝑥3 𝜕𝑓2 𝜕𝑥1 𝜕𝑓3 𝜕𝑥2 𝜕𝑓3 𝜕𝑥3 𝜕𝑓3 𝜕𝑥2 𝜕𝑥3 ⋮ ⋮ ⋮ ⋮ ⋮ ⋮ 𝜕𝑓𝑛 𝜕𝑓𝑛 [𝜕𝑥1 𝜕𝑓𝑛 𝜕𝑥2 𝜕𝑥3 F ′ (X) = J(X) = 𝜕𝑥 1 … 𝜕𝑓1 𝜕𝑥𝑛 𝜕𝑓2 x1 x2 … , where X = x3 𝜕𝑥𝑛 ⋮ … ⋮ [x n ] … ⋮ 𝜕𝑓𝑛 … 𝜕𝑥𝑛 ] … 𝜕𝑥𝑛 𝜕𝑓3 14 Convex set: Let K be a non-empty set in Rn. The set K is said to be a convex set if for all U and V which are subsets of K and for t ∈ [0, 1], then (1 - t) U + tV is again a subset of K. Fréchet-Differentiability: A mapping F : D⊆ Rn →Rn is said to be Frèchet-differentiable at X∈int(D), if there exists a linear operator F′ (X) ∈ L(Rn) such that for any h ∈ Rn, then lim ‖F(X+h)−F(X)−F′ (X)h‖ ‖h‖ h→0 = 0. Lipschitz Continuity: A mapping F: D ⊆ Rn → Rn, is said to be Lipschitz continuous on D0 ⊆ D, if there exists a constant L ≥ 0, such that, ∀ x1, x2∈D0, ‖F(x1 ) − F(x2 )‖ ≤ L‖x1 − x2 ‖ Order of Convergence: Let 𝜶 be a root of the function F(X) = 0 and suppose X (k−2), X (k−1) , X (k) and X (k+1) are the last four consecutive iterations which are very closer and closer to the true value 𝜶 of an iterative method. Then, the computational order of convergence ρ of the iterative method can be approximated by using the formula ρ≈ ln(‖X (k+1) − X (k) ‖⁄‖X (k) − X (k−1) ‖) ln(‖X (k) − X (k−1) ‖⁄‖X (k−1) − X (k−2) ‖) There are different orders of convergences. These are as defined below. A sequence X (k+1) is said to be linearly convergent if X (k+1) converges to the actual root 𝜶 with order 𝑝 = 1, for a constant 𝛿 < 1 such that ‖X (k+1) − 𝛼‖ ‖X (k+1) − 𝛼‖ lim = lim =𝛿 𝑘→∞ ‖X (k) − 𝛼‖𝑝 𝑘→∞ ‖X (k) − 𝛼‖1 A sequence X (k+1) is said to be quadratically convergent if X (k+1) converges to the actual root 𝜶 with order 𝑝 = 2 such that ‖X (k+1) − 𝛼‖ =𝛿 𝑘→∞ ‖X (k) − 𝛼‖2 lim A sequence X (k+1) is said to be super linearly convergent if ‖X (k+1) − 𝛼‖ =0 𝑘→∞ ‖X (k) − 𝛼‖2 lim 15 The higher the value of the order of convergence p is, the more rapid the convergence of the sequence is. Stopping (or Termination) Criteria: The iterative method should be stopped if the approximate solution is accurate enough. A good termination criterion is very important, because if the criterion is too weak the solution obtained may be useless, whereas if the criterion is too severe the iteration process may never stop or may cost too much work. The iteration process would be stopped when iteration error is just small enough. The difference between a computed iterate and the true solution of a linear system, measured in some vector norm. Mathematically; Let {X (k) }𝒌≥𝟎 be a sequence in Rn convergent to 𝜶, and ∈>0 be a given tolerance. Then, one of the following conditions can be used as termination (stopping) criteria: ‖F(X (k) )‖ < 𝜀 ‖X (k) − X (k−1) ‖ < 𝜀 Or, ‖𝐗 (𝐤) −𝐗 (𝐤−𝟏) ‖ ‖𝐗 (𝐤) ‖ < 𝜀, X (k) ≠ 0 The main issue of the iterative method is to check or to prove if the sequence really converges to a fixed point X ∗ . If not, we say that the method diverges. If the method diverges no matter how close the initial approximation X (0) to the fixed point X ∗ is, we say that the algorithm is unstable. Otherwise, the algorithm could be stable for smaller errors 𝑒 (𝑘) = X (k) − X ∗ but it could diverges for larger errors 𝑒 (𝑘) . Convergence of Iterative Methods: An iterative method is said to be convergent if the corresponding sequence converges for the given initial approximations. Or, a sequence is said to be convergent if it has a limit. An iterative method converges for fixed initial guesses if and only if X (𝑘) = X ∗ 𝑎𝑛𝑑 𝐹(X ∗ ) = 0. Banach Lemma: Let C∈ 𝑅 𝑛×𝑛 with ‖𝐶‖ < 1, then I + C is invertible and 1 1 ≤ ‖(I + C)−1 ‖ ≤ 1 + ‖𝐶‖ 1 − ‖𝐶‖ 16 Existence and Uniqueness of Solutions for SNLEs: It is often difficult to determine the existence or number solutions to SNLEs. For system of linear equations the number of solutions must be zero, one or infinitely many. But nonlinear equations can have any number of solutions. Thus, determining existence and uniqueness of solutions is more complicated for nonlinear system of equations than for linear equations. Although it is difficult to make any assertion about solution of nonlinear equations, there are nevertheless some useful local criteria that guarantee existence of a solution. The simplest for these is for one-dimensional problems, for which the Intermediate Value Theorem provides a sufficient condition for a solution, which says that if f is continuous on a closed interval [a, b], and c lies between f(a) and f(b), then there is a value x*∈[a,b] such that f(x*) = c. Thus, if f(a) and f(b) differ in sign, then by taking c = 0 in the theorem we can conclude that there must be a root within the interval [a,b]. Such an interval [a, b] for which the sign of f differs at its endpoints is called a bracket for a solution of the one-dimensional nonlinear equation f(x) = 0. Note: There is no simple analog for n dimensions. The nonlinear equations can have any number of solutions. It can have a simple as well as multiple roots. 4.2. Some Quadrature Formulas The following are some useful quadrature formulas (Burden and Faires, 2001; Podisuk et al., 2007), which are needed in the study of the next chapters. One point open quadrature formula (Midpoint rule) is given by 𝑏 𝑎+𝑏 ∫𝑎 𝑓(𝑥)𝑑𝑥 ≈ (𝑏 − 𝑎)𝑓 ( 2 ) (4.1) Two point closed quadrature formula (Trapezoidal rule) is also given by: 𝑏 ∫𝑎 𝑓(𝑥)𝑑𝑥 ≈ (𝑏−𝑎) 2 [𝑓(𝑎) + 𝑓(𝑏)] (4.2) A two point closed-open quadrature formula is given by: 𝑏 ∫𝑎 𝑓(𝑥)𝑑𝑥 ≈ (𝑏−𝑎) 4 𝑎+2𝑏 [𝑓(𝑎) + 3𝑓 ( 3 )] (4.3) Another two point closed-open quadrature formula is given by: 𝑏 ∫𝑎 𝑓(𝑥)𝑑𝑥 ≈ (𝑏−𝑎) 4 2𝑎+𝑏 [3𝑓 ( 3 ) + 𝑓(𝑏)] (4.4) 17 5. SINGLE STEP ITERATIVE METHODS FOR SOLVING SNLEs As it is discussed earlier in the background section, there are a number of iterative methods developed by a number of researchers to solve system of nonlinear equations. Some of these methods are methods containing a single step or they are methods containing two or more steps. Here in this chapter, the two most popular numerical iterative methods with a single step are discussed. 5.1. Newton Method Newton’s method is one of the most popular iterative methods that is used to approximate the roots of a given system of nonlinear equations. The historical road to Newton’s method is interesting in its own right, setting aside the technical details. Sir Isaac Newton did indeed play a role in its development, but there were certainly other key players. Despite the uncertainties, most experts agree that the road begins with a perturbation technique for solving scalar polynomial equations, pioneered by Francois Vieta in 1600 (Deuflhard, 2012). This method used one decimal place of the computed solution on each steps, and converged linearly (Deuflhard, 2012). In 1664 Isaac Newton learned of Vieta’s method, and by 1669 he had improved it by linearizing the resulting equations (Deuflhard, 2012). One of his first applications of this newfangled scheme was on the numerical solution of the cubic polynomial equation 𝑓(𝑥) = 𝑥 3 − 2𝑥 − 5 = 0. By using an initial guess 𝑥 (0) = 2, he created an iterative method by substituting 𝑥 (𝑘) = 𝑥 (𝑘−1) + δ and solved for δ (neglecting terms higher than the first order), and repeated the process, etc, which produced a better and better approximation to the true root (Deuflhard, 2012). Newton realized that by keeping all decimal places of the corrections, the number of accurate places would double, which means quadratic convergence. However, there is no evidence that Newton incorporated derivatives in his method. In 1690, Joseph Raphson improved Newton’s method of 1669 by making the scheme fully iterative, making computation of successive polynomials unnecessary (Deuflhard, 2012). But it was in 1740 when the derivative aspect of ”Newton’s Method” was brought to the fore by Thomas Simpson (Deuflhard, 2012). Simpson extended Newton’s formulation from 1669 by introducing derivatives, and wrote the mathematical formula for the iterative method for both a scalar nonlinear equation, and a system of two equations and two unknowns (Deuflhard, 18 2012). For a scalar equation and some initial guess 𝑥 (0) , this formulation was given by the equation, 𝑥 (𝑘) = 𝑥 (𝑘−1) − 𝑓(𝑥 (𝑘−1) ) , 𝑓 ′ (𝑥 (𝑘−1) ) where 𝑘 = 1,2, … Thus, it was Simpson who made the leap from using this procedure on scalar equations to systems of equations. Systems of equations arise so frequently in not only just pure mathematics, but also in the physical science and engineering disciplines. The physical laws that direct mathematicians’ and engineers’ research can almost always be transformed from a continuous system to a discrete system of equations, and more often than not, they are nonlinear. In fact, nonlinear systems abound in mathematical problems. This makes the phenomena being modeled more diverse and interesting, but, the price to pay is that finding solutions are nontrivial. The two main classes of methods used for solving nonlinear algebraic equations are Newton and quasiNewton method. Equivalently, it could be said that the two classes of methods are derivative and derivative-free, respectively. Assume F: D⊆Rn→Rn is r- times Fréchet differentiable function in D and has a unique solution, X (∗) ∈ Rn which is the zero of the given system of nonlinear equations, then, for any x (k+1), x (k) ∈D performing the Taylor expansion of F at point X (k) , we have: 1 ′′ (k) 2 F (X )(X (∗) − X (k) ) 2! 1 1 3 4 + F ′′′ (X (k) )(X (∗) − X (k) ) + F (iv) (X (k) )(X (∗) − X (k) ) + ⋯ 3! 4! F(X (∗) ) = F(X (k) ) + F ′ (X (k) )(X (∗) − X (k) ) + 2 ⟹ F(X (∗) ) = F(X (k) ) + F ′ (X (k) )(X (∗) − X (k) ) + O (‖X (∗) − X (k) ‖ ), where F ′ (X (k) ) is a Jacobian matrix associated with F. 2 The Newton’s method is based on a linear approximation, so that O(‖X (∗) − X (k) ‖ ) and higher order terms are neglected; i.e., if we take the first two terms of the Taylor’s series expansion we have: F(X (∗) ) ≈ F(X (k) ) + F ′ (X (k) )(X (∗) − X (k) ) Since X (∗) is the root of F, then we obtain 19 F(X (k) ) + F ′ (X (k) )(X (∗) − X (k) ) = 0 ⟹ F ′ (X (k) )(X (∗) − X (k) ) = −F(X (k) ) Multiplying both sides by the inverse of F ′ (X (k) ) if it exists, will give the following result: (F ′ (X (k) )) −1 −1 F ′ (X (k) )(X (∗) − X (k) ) = − (F ′ (X (k) )) ⟹ X (∗) − X (k) = − (F ′ (X (k) )) −1 ⟹X (∗) = X (k) − (F ′ (X (k) )) −1 F(X (k) ) F(X (k) ) F(X (k) ) −1 ⟹ x (k+1) = X (k) − F ′ (X (k) ) F(X (k) ) (5.1) where, F ′ (X k ) is a Jacobian Matrix evaluated at a point X 𝑘 , for k= 0, 1, 2, …, and take x (k+1) = X (∗) . So equation (5.1) is called Newton’s method for solving SNLEs F(X) = 0 and is generally expected to give a quadratic convergence. That means, if an initial vector X (0) is chosen such that it is sufficiently close to the true root X (∗) and J(x) is nonsingular and Lipschitz continuous for all 𝑥 close to X (∗) , then the Newton’s method enjoys a quadratic convergence, meaning the error at iteration k is proportional to the square of the error at iteration k-1 (Martinez, 2000). Quadratic Convergence Analysis of Newton’s Method: The Newton method is best known for its quadratic convergence. Toward this, we first prove a useful lemma. Lemma 1: Let F: Rn →Rn be continuously differentiable in an open convex set D⊆Rn. Suppose a constant 𝛾 exists such that ‖𝐹 ′ (𝑥) − 𝐹 ′ (𝑦)‖ ≤ 𝛾‖𝑥 − 𝑦‖ for all 𝑥, 𝑦 ∈ 𝐷. 𝛾 Then, ‖𝐹(𝑥) − 𝐹(𝑦) − 𝐹 ′ (𝑦)(𝑥 − 𝑦)‖ ≤ 2 ‖𝑥 − 𝑦‖2 Proof: By the line integration, 1 𝐹(𝑥) − 𝐹(𝑦) = ∫ 𝐹 ′ (𝑦 + 𝑡(𝑥 − 𝑦)) (𝑥 − 𝑦)𝑑𝑡 So, 𝐹(𝑥) − 𝐹(𝑦) − 𝐹 ′ (𝑦)(𝑥 − 𝑦) = 0 1 ′ ∫0 𝐹 (𝑦 + 𝑡(𝑥 − 𝑦) − 𝐹 ′ (𝑦)) (𝑥 − 𝑦)𝑑𝑡 1 ′ ⟹ ‖𝐹(𝑥) − 𝐹(𝑦) − 𝐹 (𝑦)(𝑥 − 𝑦)‖ ≤ ∫‖𝐹 ′ (𝑦 + 𝑡(𝑥 − 𝑦) − 𝐹 ′ (𝑦))‖ ‖(𝑥 − 𝑦)‖𝑑𝑡 0 20 1 𝛾 ≤ ∫0 𝛾𝑡‖𝑥 − 𝑦‖2 𝑑𝑡 = 2 ‖𝑥 − 𝑦‖2 Theorem: Let F: Rn →Rn be continuously differentiable in an open convex set D⊆Rn. Assume that ∃𝜉 ∈ Rn and 𝛽, 𝛾 > 0 such that F(𝜉) = 0 F ′ (𝜉)−1 exists ‖F ′ (𝜉)−1 ‖ ≤ 𝛽 and ‖F ′ (𝑥) − F ′ (𝑦)‖ ≤ 𝛾‖𝑥 − 𝑦‖ for 𝑥, 𝑦 in a neighborhood of 𝜉. Then ∃∈> 0 such that for all 𝑥0 ∈ (𝜉 , ∈), the sequence {𝑥 (𝑘) } defined by 𝑥 (𝑘+1) = 𝑥 (𝑘) + 𝑠𝑘 and 𝐹 ′ (𝑥 (𝑘) )𝑥 (𝑘) = −F(𝑥 (𝑘) ) is well defined and converges to 𝜉 and satisfies, ‖𝑥 (𝑘+1) − 𝜉‖ ≤ 𝛽𝛾‖𝑥 (𝑘) − 𝜉‖ 2 Proof: By continuity of F ′ , choose ∈≤ 𝑚𝑖𝑛 {𝛾, 𝑥 ∈ 𝑁(𝜉, 1 2𝛽𝛾 } so that 𝐹 ′ (𝑥) is nonsingular for all ∈). For 𝑘 = 0, ‖‖𝑥 (0) − 𝜉‖‖ <∈. 1 So, ‖F ′ (𝜉)−1 (𝐹 ′ (𝑥 (0) ) − F ′ (𝜉))‖ ≤ ‖F ′ (𝜉)−1 ‖‖𝐹 ′ (𝑥 (0) ) − F ′ (𝜉)‖ ≤ 𝛽𝛾‖𝑥 (0) − 𝜉‖ ≤ 2. −1 By Banach Lemma, ‖F ′ (𝜉)−1 ‖ = ‖[F ′ (𝜉)−1 + (𝐹 ′ (𝑥 (0) ) − F ′ (𝜉))] ‖ ≤ ‖F ′ (𝜉)−1 ‖ ≤ 2‖F ′ (𝜉)−1 ‖ ≤ 2𝛽 1 − ‖F ′ (𝜉)+−1 (𝐹 ′ (𝑥) − F ′ (𝜉))‖ Now, 𝑥1 − 𝜉 = 𝑥 0 − 𝜉 − F ′ (𝑥 (0) )−1 𝐹 ′ (𝑥 (0) ) = 𝑥 0 − 𝜉 − F ′ (𝑥 (0) )−1 (𝐹(𝑥 (0) ) − F(𝜉)) = F ′ (𝑥 (0) )−1 [F(𝜉) − 𝐹(𝑥 (0) ) − 𝐹 ′ (𝑥 (0) )(𝜉 − 𝑥 (0) )] 𝛾 So, ‖𝑥1 − 𝜉‖ ≤ ‖F ′ (𝑥 (0) )−1 ‖‖F(𝜉) − 𝐹(𝑥 (0) ) − 𝐹 ′ (𝑥 (0) )(𝜉 − 𝑥 (0) )‖ ≤ 2𝛽 2 ‖𝜉 − 𝑥 (0) ‖ 2 = 𝛽𝛾‖𝜉 − 𝑥 (0) ‖ (by Banach Lemma) ≤ 𝛽𝛾𝜖‖𝑥 (0) − 𝜉‖ ≤ The proof is completed by induction. 1 (0) ‖𝑥 − 𝜉‖ ≤ 𝜖 2 2 21 Note: The above theorem shows that the Newton’s method converges quadratically if F ′ (𝜉) is non-singular and if the starting point is very close to 𝜉. Advantages and Disadvantages of Newton’s method: One of the advantages of Newton’s method is that it is not too complicated. The major disadvantage associated with this method is that Jacobian matrix as well as its inversion has to be calculated for each iteration. Calculating both Jacobian matrix and its inverse can be quite time consuming depending on the size of SNLEs. Another problem that we may be challenged with when using Newton’s method fails to converge. If Newton’s method fails to converge, this will result in an oscillation between points. 5.2. Broyden’s Method Charles Broyden was a Physicist working with nonlinear reactor models for the English electric industry company in 1960’s in Leicester (Griewank, 2012). He had to solve a problem involving non-linear algebraic equations. Broyden was well aware of the shortcomings of the Newton’s method and thought of the way to overcome them. When trying to solve nonlinear systems that arose from the discretization of these models, he realized that instead of repeatedly evaluating and factoring the Jacobian matrix in Newton’s method, he could use secant information (i.e. function value diff erences and the previous solution information) to directly compute an approximation matrix B (k) to the Jacobian (Griewank, 2012). In the onedimensional case, the problem is trivial because a unique approximation B(k) can be computed by dividing the function diff erence by the solution diff erence. Thus, letting 𝑦 (𝑘) = F(X (k) ) − F(X (k−1) ) and 𝛿 (𝑘) = X (k) − X (k−1) , B(k) = 𝑦 (𝑘) ≈ F ′(X (k) ) (𝑘) 𝛿 For multi-dimensional system of nonlinear equations, Broyden realized that he doesn’t necessarily need to work with the true inverse of Jacobian matrix, but with a suitable approximation matrix B (k) . His method is the generalization of the secant method to multiple dimensions. Newton's method for solving the equation F(X) = 0 uses the Jacobian matrix at every iteration and computing this Jacobian is a difficult and expensive operation. So the idea 22 behind Broyden's method is to compute the whole Jacobian only at the first iteration, and to do a rank-one update at the other iterations. His method suggests that the Jacobian approximation from the previous iteration, B(k−1), could be updated with a certain rank-one update. This update would satisfy not only the secant condition, but also the condition that B(k) w = 𝑇 B(k−1) w for all directions w ∈ 𝑅 𝑛 orthogonal to 𝛿 (𝑘) (i.e.(𝑤 (𝑘) ) 𝛿 (𝑘) = 0) (Griewank, 2012). The rank-one update is the case when an n×n matrix B has the form, B = A + u ⊗ v, where u and v are given vectors in Rn, and u ⊗ v is the outer product of u and v, defined by u1 v1 u1 v2 ⋯ u1 vn u u2 v2 ⋯ u2 vn 2 v1 u ⊗ v = uv T = [ ⋮ ⋮ ⋮ ] ⋱ un v1 un v2 ⋯ un vn This modification of A to obtain B is called a rank-one update. This is because u ⊗ v has rank one, since every column of u ⊗ v is a scalar multiple of u. Now to gain a sense of how this method is developed, consider a linear approximation of F(X) about some X (k) . Suppose F: D⊆Rn→Rn is r- times Fréchet differentiable function in D and has a unique solution, X (∗) ∈ Rn which is the zero of the given system of nonlinear equations, then, for any x (k+1), x (k) ∈D after performing the Taylor expansion of F at some point X (k) , we have the following condition. F(X (k−1) ) ≈ F(X (k) ) + B(k) (X (k−1) − X (k) ), where B(k) = F′(X (k) ) and k = 1, 2, . . . ⟹ F(X (k) ) − F(X (k−1) ) = B(k) (X (k) − X (k−1) ) (5.2) Let 𝛿 (𝑘) = X (k) − X (k−1) and 𝑦 (𝑘) = F(X (k) ) − F(X (k−1) ). Now equation (5.2) becomes, B(k) 𝛿 (𝑘) = 𝑦 (𝑘) (5.3) Equation (5.3) (Sauer, 2006) is called the secant relation. In order to determine a unique B (k) , Broyden pointed that since there is no new information obtained from the orthogonal complement of 𝛿 (𝑘) , then there is no reason why B(k−1) would change in a direction w, 𝑇 where (𝛿 (𝑘) ) (𝑤) = 0. So rather than computing B(k) from the scratch, the previous 23 approximation B (k−1) should be updated such that both the secant condition and equation (5.4) is satisfied. B(k) w = B(k−1) w (5.4) An approximate Jacobian satisfying both equations (5.3) and (5.4) is given by 𝑇 B (k) =B (k−1) + ( 𝑦 (𝑘) −B(k−1) 𝛿 (𝑘−1) ) (𝛿 (𝑘−1) ) (5.5) 𝑇 (𝛿 (𝑘−1) ) 𝛿 (𝑘−1) Finally, the iterative solution for Broyden’s method becomes: X (k+1) = X (k) − (B(k) ) −1 (F(X (k) )) (5.8) Broyden’s method (Griewank, (1986)) converges super linearly. Superlinear Convergence Analysis of Broyden’s Method To prove the superlinear convergence of this method, let’s first see the following lemmas. Lemma 2: Let F:Rn →Rn be continuously differentiable in an open convex set D⊆Rn. Suppose a constant 𝛾 exists such that ‖𝐹 ′ (𝑥) − 𝐹 ′ (𝑦)‖ ≤ 𝛾‖𝑥 − 𝑦‖ for all 𝑥, 𝑦 ∈ 𝐷. Then it holds that for all 𝑥, 𝑦, 𝜉 ∈ 𝐷, 𝛾 ‖𝐹(𝑥) − 𝐹(𝑦) − 𝐹 ′ (𝜉)(𝑥 − 𝑦)‖ ≤ (‖𝑥 − 𝜉‖ + ‖𝑦 − 𝜉‖)‖𝑥 − 𝑦‖ 2 Proof: By the line integral 1 ‖𝐹 ′ (𝑥) − 𝐹 ′ (𝑦) − 𝐹 ′ (𝜉)(𝑥 − 𝑦)‖ = ‖∫‖𝐹 ′ (𝑦 + 𝑡(𝑥 − 𝑦) − 𝐹 ′ (𝜉))‖ ‖(𝑥 − 𝑦)‖𝑑𝑡‖ 0 1 1 ≤ 𝛾‖𝑥 − 𝑦‖ ∫‖𝑦 + 𝑡(𝑥 − 𝑦) − 𝜉‖ 𝑑𝑡 ≤ 𝛾‖(𝑥 − 𝑦)‖ ∫{𝑡‖𝑥 − 𝜉‖ + (1 − 𝑡)‖𝑦 − 𝜉‖}𝑑𝑡 0 ≤ 0 𝛾 (‖𝑥 − 𝜉‖ + ‖𝑦 − 𝜉‖)‖𝑥 − 𝑦‖ 2 Lemma 3: Let F: Rn →Rn be continuously differentiable in an open convex set D⊆Rn.Suppose a constant 𝛾 exists such that ‖𝐹 ′ (𝑥) − 𝐹 ′ (𝑦)‖ ≤ 𝛾‖𝑥 − 𝑦‖ for all 𝑥, 𝑦 ∈ 𝐷. The for X (k+1) , X (k) ∈ D, holds that ‖𝐵 (𝑘+1) − 𝐹′(𝜉)‖ ≤ ‖𝐵 (𝑘) − 𝐹′(𝜉)‖ + ‖X (k+1) − 𝜉‖ + ‖X (k) − 𝜉‖ Proof: By definition, 24 𝐵 (𝑘+1) =𝐵 (𝑘) −𝐹 ′ (𝜉) =𝐵 (𝑘) −𝐹 ′ (𝜉) (∆F(X (k) ) − 𝐵 (𝑘) (𝑠𝑘 ))(𝑠𝑘 )T + (𝑠𝑘 )T (𝑠𝑘 ) (∆F(X (𝑠𝑘 )(𝑠𝑘 )T (𝑠𝑘 )(𝑠𝑘 )T (𝐼 − ) − 𝐹′(𝜉) (𝐼 − ) + (𝑠𝑘 )T (𝑠𝑘 ) (𝑠𝑘 )T 𝑠𝑘 (k) ) − 𝐹′(𝜉)) (𝑠𝑘 )(𝑠𝑘 )T (𝑠𝑘 )T (𝑠𝑘 ) (𝑠 )(𝑠 )T (∆F(X(k) )−𝐹′ (𝜉))(𝑠𝑘 )(𝑠𝑘 )T Taking norm, ‖𝐵 (𝑘+1) − 𝐹 ′ (𝜉)‖ ≤ ‖𝐵 (𝑘) − 𝐹 ′ (𝜉)‖ ‖𝐼 − (𝑠𝑘 )T (𝑠𝑘 )‖ + ‖ 𝑘 𝑘 (𝑠𝑘 )T (𝑠𝑘 ) ‖ (𝑠 )(𝑠 )T But, ‖𝐼 − (𝑠𝑘 )T 𝑘 ‖ ≤ 1 𝑘 (𝑠𝑘 ) Therefore, the 3rd term is estimated by, (∆F(X (k) ) − 𝐹 ′ (𝜉)) (𝑠𝑘 )(𝑠𝑘 )T {[F (X(k+1) − F(X(k) ) − F′ (𝜉)(X(k+1) − X(k) ))]} (𝑠𝑘 )T ‖ ‖=‖ ‖ (𝑠𝑘 )T (𝑠𝑘 ) (𝑠𝑘 )T (𝑠𝑘 ) 𝛾 ≤ 2 (‖X (k+1) − 𝜉‖ + ‖X (k) − 𝜉‖) (By lemma 2) Theorem: Let F: Rn →Rn be continuously differentiable in an open convex set D⊆Rn. Suppose ∃𝜉 ∈Rn, 𝛽, 𝛾 > 0 such that F(𝜉) = 0 F ′ (𝜉)−1 exists ‖F ′ (𝜉)−1 ‖ ≤ 𝛽 and ‖F ′ (𝑥) − F ′ (𝑦)‖ ≤ 𝛾‖𝑥 − 𝑦‖ for 𝑥, 𝑦 in a neighborhood of 𝜉. Then ∃𝛿1 , 𝛿2 > 0 such that, if ‖X (0) − 𝜉‖ < 𝛿1 and ‖𝐵 (0) − 𝐹 ′ (𝜉)‖ < 𝛿2 , then the Broyden’s method is well defined, converges to 𝜉, and satisfies ‖X (k+1) − 𝜉‖ ≤ 𝐶𝑘 ‖X (k) − 𝜉‖, With lim Ck = 0 (Superlinear convergence) k→∞ 1 2𝛿 Proof: Choose 𝛿2 ≤ 6𝛽 and 𝛿1 ≤ 5𝛾 25 ‖F ′ (𝜉)−1 𝐵 (0) − 𝐼‖ ≤ 𝛽𝛿2 ≤ By Banach Lemma (B (0) ) −1 1 6 exists. So 𝑥 (1) can be defined furthermore, −1 ‖F′ (𝜉)−1 ‖ −1 𝛽 ‖(B(0) ) ‖ = ‖F ′ (𝜉) + (B(0) − F ′ (𝜉)) ‖ ≤ 1−‖F′ (𝜉)−1 ‖‖B(0)−F′ (𝜉)‖ ≤ 1−𝛽𝛿 . 2 ‖𝜖1 ‖ = ‖x(1) − 𝜉‖ = ‖x(0) − (B(0) ) −1 (F(x(0) ) − F(𝜉)) − 𝜉‖ = ‖−(B(0) )−1 [𝐹(x (0) ) − F(𝜉) − B (0) (x (0) − 𝜉)]‖ = ‖(B(0) )−1 [𝐹(x (0) ) − F(𝜉) − F′ (𝜉)(x (0) − 𝜉) + (F′ (𝜉) − B (0) )(x (0) − 𝜉)]‖ ≤ 𝛽 𝛾 𝛽 [𝛾, 𝑣𝑜𝑒𝑟2𝛿1 + 𝛿2 ]‖𝜖0 ‖ [ ‖𝜖0 ‖2 + 𝛿2 ‖𝜖0 ‖] ≤ 1 − 𝛽𝛿2 2 1 − 𝛽𝛿2 1 𝛽 6 6 1 ≤ ( 𝛿2 ) ‖𝜖0 ‖ ≤ 6 1 ‖𝜖0 ‖ ≤ ‖𝜖0 ‖ 1 − 𝛽𝛿2 5 2 1−65 From lemma 2, 𝛾 ‖𝐵 (1) − 𝐹 ′ (𝜉)‖ ≤ ‖𝐵 (0) − 𝐹 ′ (𝜉)‖ + (‖x(1) − 𝜉‖ + ‖x(0) − 𝜉‖) 2 𝛾 3 𝛾 32 3 3 ≤ 𝛿2 + ( ‖𝜖0 ‖) ≤ 𝛿2 (1 + ) = (1 + ) 𝛿2 ≤ 𝛿2 2 2 225 10 2 1 Thus, ‖F ′ (𝜉)−1 𝐵 (0) − 𝐼‖ ≤ 2𝛽𝛿2 ≤ 3. By Banach Lemma (B (0) ) −1 ‖(B (0) ) ‖ ≤ −1 exists, then ‖F′ (𝜉)−1 ‖ 1−‖F′ (𝜉)−1 ‖‖B(1)−F′ (𝜉)‖ ≤ 𝛽 1−2𝛽𝛿2 3 ≤ 𝛽 2 The following estimation can now be made, ‖𝜖2 ‖ = ‖x(2) − 𝜉‖ = ‖x(1) − (B (1) ) −1 (F(x(1) ) − F(𝜉)) − 𝜉‖ = ‖−(B(1) )−1 [𝐹(x (1) ) − F(𝜉) − B (1) ϵ1 ]‖ = ‖(B(1) )−1 [𝐹(x (1) ) − F(𝜉) − F′ (𝜉)ϵ1 + (F′ (𝜉) − B (1) )ϵ1 ]‖ = 3𝛽 𝛾 3 3𝛽 𝛾 𝛿1 3 [‖𝜖1 ‖2 + 𝛿2 ‖ϵ1 ‖] = [ + 𝛿 ] ‖ϵ1 ‖ 2 2 2 2 2 2 2 2 26 ≤ 3𝛽𝛿2 𝛾 1 2 3 1 16 1 ‖ϵ0 ‖ ≤ ‖ϵ0 ‖ [ + ] ‖ϵ1 ‖ ≤ 2 2 2 5𝛾 2 4 10 2 Continuing 𝛾 ≤ ‖B(2) − 𝐹′(𝜉)‖ ≤ ‖B1 − 𝐹′(𝜉)‖ + (‖ϵ1 ‖ + ‖ϵ2 ‖) 2 ≤ 13𝛿2 𝛾 3 3 𝛾 31 2 3 1 3 1 2 + ( ‖ϵ1 ‖) ≤ 𝛿2 (1 + + ) ≤ 𝛿2 (1 + + ) ≤ 𝛿2 (2 − ( ) ) ≤ 2𝛿2 10 2 2 10 2 2 2 5𝛾 10 2 10 2 The proof is complete by mathematical induction. Advantages and Disadvantages of Broyden’s method: The main advantage of Broyden’s method is the reduction of computations; more specifically, the way the inverse of the approximation matrix, (B (k) ) previous iteration, (B (k−1) ) −1 −1 can be computed directly from the reduces the number of computations needed for this method in comparison to NM. The disadvantage of this method is that it does not converge quadratically. This means that more iteration may be needed to reach the solution, when compared to the number of iterations Newton method requires. Another disadvantage of this method is that it is not self-correcting. This means that in contrast to NM, it does not correct itself round of errors with consecutive iterations. This may cause only a slight inaccuracy in the iteration compared to NM, but the final iteration result will be the same. 27 6. TWO ITERATIVE METHODS WITH TWO STEPS FOR SOLVING SNLEs In this chapter, two iterative methods with two-steps each, commonly known as predictorcorrector methods for solving SNLEs F(X) = 0 are derived and analyzed. The closed-open quadrature rules are used to derive the correctors but the predictor for both cases is the Newton’s method. It is shown that both methods are cubic convergent. The Taylor series technique is used for the proof of convergence analysis of the two methods. To illustrate the performance of the two methods some numerical examples are solved. The comparison of the two methods with Newton’s method and Broyden’s method is also given. 6.1. Predictor-corrector Methods 1 and 2 Consider the system of nonlinear equations F(X) = 0 (6.1) where X = (x1 , x2 , x3 , … , xn )𝑇 ∈ Rn . Suppose that D⊆Rn be open and convex set in Rn. Let F: D⊆ Rn → Rn be r- times Fréchet differentiable function in D and 𝛽 ∈ Rn be a zero of the given system of nonlinear equations. For any X, X (k) ∈ D, the Taylor formula for the function F can be written as: 1 ′′ (k) 2 F (X )(X − X (k) ) 2! 1 1 1 3 4 5 + F ′′′ (X (k) )(X − X (k) ) + F (iv) (X (k) )(X − X (k) ) + F (v) (X (k) )(X − X (k) ) 3! 4! 5! 1 (r−1) +⋯+ F (r−1) (X (k) )(X − X (k) ) (r − 1)! F(X) = F(X (k) ) + F ′ (X (k) )(X − X (k) ) + 1 (1−t)(r−1) + ∫0 F (r) (X (k) + t(X − X (k) ))(X − X (k) ) (r) dt (6.2) F(X) = F(X (k) ) + ∫0 F ′ (X (k) + t(X − X (k) )) (X − X (k) )𝑑𝑡 (6.3) (r−1)! where 𝑡 ∈ [0,1]. If we take 𝑟 = 1, then equation (6.2) becomes 1 If we approximate integral in equation (6.3) by using closed-open quadrature formula based on equation (4.3), we have 28 1 1 ∫0 F ′ (X (k) + t(X − X (k) )) (X − X (k) )dt ≈ 4 [F ′ (X (k) ) + 3F ′ ( X(k) +2X 3 )] (X − X (k) ) (6.4) Substituting equation (6.4) into equation (6.3), we have: X(k) +2X 1 F(X) = F(X (k) ) + 4 [F ′ (X (k) ) + 3F ′ ( 3 )] (X − X (k) ) (6.5) From equation (6.1), F(X) = 0 based on which equation (6.5) becomes 1 ′ (k) X (k) + 2X (k) ′ 0 = F(X ) + [F (X ) + 3F ( )] (X − X (k) ) 4 3 Subtracting F(X (k) ) from both sides, we have X(k) +2X 1 [F ′ (X (k) ) + 3F ′ ( 4 3 )] (X − X (k) ) = − F(X (k) ) X(k) +2X ⟹ [F ′ (X (k) ) + 3F ′ ( 3 )] (X − X (k) ) = −4F(X (k) ) If we multiply both sides by the inverse of F ′ (X (k) ) + 3F ′ ( −1 ′ [F (X (k) X (k) + 2X ) + 3F ( )] 3 ′ ′ = −4 [F (X (k) X(k) +2X 3 ), it becomes X (k) + 2X [F ′ (X (k) ) + 3F ′ ( )] (X − X (k) ) 3 X (k) + 2X ) + 3F ( )] 3 −1 F(X (k) ) ′ −1 ⟹X−X (k) ′ = −4 [F (X (k) X (k) + 2X ) + 3F ( )] 3 ′ X(k) +2X ⟹ X = X (k) − 4 [F ′ (X (k) ) + 3F ′ ( 3 −1 )] F(X (k) ) F(X (k) ) (6.6) Based on equation (6.6), we have the following algorithm. Algorithm 6.1: For a given initial approximation value X (0) , then the (𝑘 + 1)𝑡ℎ approximate value, X = X (k+1) of SNLEs can be computed by using the iterative scheme: X(k) +2X(k+1) X (k+1) = X (k) − 4 [F ′ (X (k) ) + 3F ′ ( 3 −1 )] F(X (k) ), where k = 0, 1, 2, … This algorithm is an implicit type method. To solve the problems of SNLEs by using algorithm 6.1, we have to find the solution of the given problem implicitly, which by itself is a difficult problem. To avoid such difficulties we use the predictor and corrector concepts. Therefore, we use the Newton’s method as a predictor and the method in algorithm 6.1 as a corrector. This gives the following predictor – corrector method for solving SNLEs. Algorithm 6.2: For a given initial value X (0) the approximate solution X (k+1) can be computed by the following iterative schemes. Predictor step: 29 Y (k) = X(k) − (F′ (X(k) )) −1 F(X(k) ), Corrector step: X (k+1) = X (k) − 4 × [F ′ (X (k) ) + 3F ′ ( X(k) +2 𝐘(𝐤) 3 −1 F(X (k) ), k = 0,1,2, … )] Algorithm 6.2 is one of the four methods on which this project paper focused to solve SNLEs. This method is referred to as predictor-corrector method-1 (PCM-1). Again approximating the integral given in (6.3) by another Closed-open quadrature formula (based on (4.16)), we have: 1 2X(k) +X 1 ∫0 F ′ (X (k) + t(X − X (k) )) (X − X (k) ) 𝑑𝑡 ≈ 4 [3F ′ ( 3 ) + F ′ (X)] (X − X (k) ) (6.7) Replacing (6.7) in (6.3), we obtain 2X(k) +X 1 F(X) = F(X (k) ) + 4 [3F ′ ( 3 ) + F ′ (X)] (X − X (k) ). (6.8) It is clear that F(X) = 0 from equation (6.1). Hence, equation (6.8) becomes: 1 2X (k) + X 0 = F(X (k) ) + [3F ′ ( ) + F ′ (X (k) )] (X − X (k) ) 4 3 Now, this equation can be written as 1 2X (k) + X ′ [3F ( ) + F ′ (X)] (X − X (k) ) = −F(X (k) ) 4 3 2X(k) +X ⟹[3F ′ ( 3 ) + F ′ (X)] (X − X (k) ) = −4F(X (k) ) Multiplying both parts of this equation by the inverse of [3F ′ ( −1 2X (k) + X [3F ( ) + F ′ (X)] 3 ′ 2X(k) +X 3 ) + F ′ (X)], it becomes 2X (k) + X [3F ′ ( ) + F ′ (X)] (X − X (k) ) 3 −1 2X (k) + X = −4 [3F ( ) + F ′ (X)] 3 ′ 2X(k) +X ⟹ X − X(k) = −4 [3F′ ( ⟹ X = X (k) − 4 [3F ′ ( 3 2X(k) +X 3 F(X (k) ) ) + F′ (X)] ) + F ′ (X)] −1 −1 F(X(k) ) F(X (k) ) (6.9) From this relation, we can arrive at the following two-step iterative scheme for solving SNLEs. 30 Algorithm 6.3: The approximate solution X = X (k+1) is calculated by using the following iterative schemes if the initial approximation value for the problem is given by X (0) . −1 X (k+1) =X (k) 2X (k) + X (k+1) − 4 [3F ( ) + F ′ (X (k+1) )] 3 ′ F(X (k) ), 𝑘 = 0, 1,2, … Algorithm 6.3 is also implicit type method. We remove the implicit problem by using NM as a predictor step in predictor-corrector method for solving SNLEs F(X) = 0. Algorithm 6.4: For a given initial value, X (0) , the approximate solution, X (k+1), can be computed by using the following iterative procedure. Predictor step: Y (k) = X(k) − (F′ (X(k) )) −1 F(X(k) ) Corrector step −1 X (k+1) 2X (k) + Y (k) (k) ′ = X − 4 [3F ( ) + F ′ (Y (𝐤) )] 3 F(X (k) ), 𝑘 = 0, 1,2, … Algorithm 6.4 is the second iterative method on which this paper focused on to solve system of nonlinear equations. We refer to it as predictor-corrector method-2 (PCM-2). 6.2. Convergence Analysis of the Two Methods Here the order of convergence for both methods given in algorithm 6.2 (PCM-1) and algorithm 6.4 (PCM-2) are shown by proving theorems 6.1 and 6.2. 6.2.1. Proof of cubic order of convergence for algorithm 6.2 Theorem 6.1 (Noor and Waseem, 2008): Let F: D⊆ Rn → Rn , be r - times Fréchet differentiable function in an open and convex set D⊆ Rn containing the root 𝜶 of F(X) = 0. Suppose that F ′ (X (k) ) is continuous and nonsingular in D. If the initial approximate value X (𝟎) is sufficiently close to the exact root 𝜶, then the iterative method defined by Algorithm 6.2 converges to 𝜶 with cubic order of convergence and satisfies the error equation X(k) +2𝐘(𝐤) )] e(k+1) 3 [F ′ (X (k) ) + 3F ′ ( −1 = [F ′′ (X (k) ) (F ′ (X (k) )) 3 4 F ′′ (X (k) ] (e(k) ) +O(‖e(k) ‖ ) (6.10) 31 Proof: The proof is adopted from Noor and Waseem, 2008. From Algorithm 6.2, the corrector method is given by the formula X (k+1) =X (k) ′ − 4 [F (X (k) )+ But, Y (k) = X (k) − (F ′ (X (k) )) X(k) +2 𝐘(𝐤) 3F ′ ( 3 −1 −1 F(X (k) ) )] (6.11) F(X (k) ) (6.12) e(k) = X (k) − 𝜶 Defining, ⟹X (k) = e(k) + 𝜶 (6.13) But from the equation, X (k+1) =X (k) ′ − 4 [F (X (k) )+ X(k) +2𝐘(𝐤) 3F ′ ( )] 3 X(k) +2𝐘(𝐤) X (k+1) − X (k) = −4 [F ′ (X (k) ) + 3F ′ ( 3 −1 F(X (k) ) we have: −1 F(X (k) ) )] Using (6.13) in the place of X (k) , it becomes ⟹X (k+1) ⟹(X − (e (k+1) (k) ′ + 𝜶) = −4 [F (X − 𝜶) − e (k) (k) ′ = −4 [F (X )+ (k) X(k) +2𝐘(𝐤) 3F ′ ( )] 3 )+ −1 F(X (k) ) X(k) +2𝐘(𝐤) 3F ′ ( )] 3 −1 F(X (k) ) If e(k) = X (k) − 𝜶, then X (k+1) − 𝜶 = e(k+1) . So the equation reduces to X(k) +2𝐘 (𝐤) ⟹e(k+1) − e(k) = −4 [F ′ (X (k) ) + 3F ′ ( 3 −1 )] F(X (k) ) (6.14) X(k) +2𝐘 (𝐤) If we pre-multiply both parts of the eq’n (6.14) by F ′ (X (k) ) + 3F ′ ( [F ′ (X (k) ) + 3F ′ ( 3 ), we obtain X (k) + 2Y (k) )] (e(k+1) − e(k) ) 3 −1 ′ = −4 [F (X ′ ⟹ [F (X (k) (k) X (k) + 2𝐘 (𝐤) X (k) + 2𝐘 (𝐤) ) + 3F ( )] [F ′ (X (k) ) + 3F ′ ( )] 3 3 ′ F(X (k) ) X (k) + 2Y (k) ) + 3F ( )] (e(k+1) − e(k) ) = −4F(X (k) ) 3 ′ X(k) +2Y(k) ⟹[F ′ (X (k) ) + 3F ′ ( 3 X(k) +2Y(k) )] e(k+1) = [F ′ (X (k) ) + 3F ′ ( 3 )] e(k) − 4F(X (k) ) (6.15) Using the equation (6.2) with X = α, F(X) = F(X (k) ) + F ′ (X (k) )(X − X (k) ) + 1 ′′ (k) 1 2 3 4 F (X )(X − X (k) ) + F ′′′ (X (k) )(X − X (k) ) + O (‖X − X (k) ‖ ) 2! 3! 32 4 where, O (‖X − X (k) ‖ ) is a truncation error with order four, we obtain F(𝛂) = F(X (k) ) + F ′ (X (k) )(𝛂 − X (k) ) + 1 ′′ (k) 1 2 3 F (X )(𝛂 − X (k) ) + F ′′′ (X (k) )(𝛂 − X (k) ) 2! 3! 4 +O (‖𝛂 − X (k) ‖ ) Since 𝛂 is the root of F, then F(𝛂) = 0. 2 1 ⟹0 = F(X (k) ) + F ′ (X (k) )(𝛂 − X (k) ) + 2! F ′′ (X (k) )(𝛂 − X (k) ) + 3 1 F ′′′ (X (k) )(𝛂 − X (k) ) + 3! 4 O (‖𝛂 − X (k) ‖ ) From the definition given in (6.13) above, we know that e(k) = X (k) − 𝜶 ⟹ −e(k) = 𝜶 − X (k) Now substituting −e(k) instead of 𝜶 − X (k) in the above equation, we get 0 = F(X (k) ) + F ′ (X (k) )(−e(k) ) + 1 ′′ (k) 1 2 3 F (X )(−e(k) ) + F ′′′ (X (k) )(−e(k) ) + O(‖−e(k) ‖4 ) 2! 3! ⟹ 0 = F(X (k) ) − F ′ (X (k) )(e(k) ) + 1 ′′ (k) (k) 2 1 3 4 F (X )(e ) − F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) 2! 3! 2 1 ⟹ F(X (k) ) = F ′ (X (k) )(e(k) ) − 2! F ′′ (X (k) )(e(k) ) + 1 3! 3 4 F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) (6.16) −1 Now multiplying equation (6.16) with (F ′ (X (k) )) the product becomes as follows. (F ′ (X (k) )) −1 F(X (k) ) −1 = (F ′ (X (k) )) [F ′ (X (k) )(e(k) ) − −1 ⟹(F ′ (X (k) )) + 1 ′′ (k) (k) 2 1 3 4 F (X )(e ) + F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )] 2! 3! −1 F(X (k) ) = (F ′ (X (k) )) −1 1 F ′ (X (k) )(e(k) ) − 2! (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 2 1 ′ (k) −1 ′′′ (k) (k) 3 4 (F (X )) F (X )(e ) + O (‖e(k) ‖ ) 3! (F ′ (X (k) )) −1 −1 1 −1 1 F(X (k) ) = (e(k) ) − 2! (F ′ (X (k) )) + 3! (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 3 X(k) +2𝐘 (𝐤) 3 2 4 + O (‖e(k) ‖ ) Now applying a Taylor series in (6.2) for F ′ ( F′ ( F ′′ (X (k) )(e(k) ) X(k) +2Y(k) X(k) +2𝐘 (𝐤) ) = F ′ (X (k) ) + F ′′ (X (k) ) ( 3 3 (6.17a) ) at the point X (k) , we have 1 X(k) +2𝐘 (𝐤) − X (k) ) + 2! F ′′′ (X (k) ) ( 3 − X (k) ) 2 33 3 4 1 X (k) + 2Y (k) X (k) + 2Y (k) + F (iv) (X (k) ) ( − X (k) ) + O (‖ − X (k) ‖ ) 3! 3 3 2 ′ = F (X (k) ′′ ) + F (X (k) 2 1 2 ) ( (−X (k) + Y (k) )) + F ′′′ (X (k) ) ( (−X (k) + Y (k) )) 3 2! 3 3 4 1 (iv) (k) 2 2 (k) (k) (k) (k) + F (X ) ( (−X + Y )) + O (‖ (−X + Y )‖ ) 3! 3 3 F′ ( X(k) +2𝐘(𝐤) 3 2 1 4 ) = F ′ (X (k) ) + 3 F ′′ (X (k) ) ((Y (k) − X (k) )) + 2! × 9 F ′′′ (X (k) )(Y (k) − X (k) ) + 1 3! × 8 27 3 4 F (iv) (X (k) )(Y (k) − X (k) ) + O (‖e(k) ‖ ) 2 (6.17b) It is clear that we use Y (k) instead of X (k+1) to separate X (k+1) of a predictor step in algorithms (6.2) and (6.4) from that of the corrector step. So we can write X (k+1) − X (k) instead of Y (k) − X (k) for the following case such that −1 X (k+1) = X (k) − (F ′ (X (k) )) F ′ (X (k) ), or −1 X (k+1) − X (k) = − (F ′ (X (k) )) F ′ (X (k) ) (6.18) Now substituting (6.18) in eq’n (6.17b), we obtain F′ ( X(k) +2𝐘 (𝐤) 3 −1 2 ) = F ′ (X (k) ) + 3 F ′′ (X (k) ) (− (F ′ (X (k) )) −1 4 + F ′′′ (X (k) ) (− (F ′ (X (k) )) F ′ (X (k) )) 18 F ′ (X (k) )) 2 3 −1 8 (iv) (k) 4 + F (X ) (− (F ′ (X (k) )) F ′ (X (k) )) + O (‖e(k) ‖ ) 162 ′ =F (X (k) 2 ′′ ) − 3 F (X (k) ′ ) ((F (X (k) −1 )) ′ F (X (k) 4 ′′′ )) + 18 F (X (k) ′ ) ((F (X (k) )) −1 2 ′ F (X (k) )) 3 −1 8 (iv) (k) 4 − F (X ) ((F ′ (X (k) )) F ′ (X (k) )) + O (‖e(k) ‖ ) 162 −1 X (k) + 2Y (k) 2 F ( ) = F ′ (X (k) ) − F ′′ (X (k) ) ((F ′ (X (k) )) F ′ (X (k) )) 3 3 ′ 2 ′′′ + F (X 9 (k) ′ ) ((F (X (k) )) −1 2 ′ F (X (k) )) − 4 81 F (𝐢𝐯) (X (k) ′ ) ((F (X Now using equation (6.17a) in (6.19), we have (k) −1 )) 3 ′ (k) 4 F (X )) + O (‖e(k) ‖ ) (6.19) 34 X(k) +2𝐘 (𝐤) F′ ( 3 2 ) = F ′ (X (k) ) − 3 F ′′ (X (k) ) ((e(k) ) − −1 3 1 (k) ( ) ′ ( F ( X )) F′′′ (X k )(e(k) ) 3! 1 ( ) (F′ (X k )) 2! −1 2 F′′ (X(k) )(e(k) ) + 4 + O (‖e(k) ‖ )) −1 −1 2 1 1 2 3 4 + F ′′′ (X (k) ) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) ) 9 2! 3! − −1 4 (iv) (k) 1 2 F (X ) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 81 2! 3 −1 1 3 4 4 + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )) + O (‖e(k) ‖ ) 3! 2 −1 1 = F ′ (X (k) ) − 3 F ′′ (X (k) )(e(k) ) + 3 F ′′ (X (k) ) (F ′ (X (k) )) F′′(X (k) )(e(k) ) 2 −1 1 3 4 − F ′′ (X (k) ) (F′(X (k) )) F′′′(X (k) )(e(k) ) + O (‖e(k) ‖ ) 9 −1 2 1 2 2 + F ′′′ (X (k) ) [(e(k) ) − (e(k) ) (F ′ (X (k) )) F′′(X (k) )(e(k) ) 9 2! + −1 1 (k) 3 4 (e ) (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) 3! 2 −1 1 ′ (k) −1 1 3 2 (F (X )) F′′(X (k) )(e(k) ) + ((F ′ (X (k) )) F′′(X (k) )(e(k) ) ) 2! 4 −1 −1 1 2 3 − [(F ′ (X (k) )) F ′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] 12 −1 1 4 + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 3! −1 −1 1 3 2 − [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′ (X (k) )(e(k) ) ] 12 + + − 2 −1 1 3 4 [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] + O (‖e(k) ‖ )] 36 −1 −1 4 (iv) (k) 1 1 2 3 F (X ) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 81 2! 3! 4 3 4 + O (‖e(k) ‖ )) + O (‖e(k) ‖ ) X(k) +2𝐘 (𝐤) ⟹ F′ ( 3 2 1 −1 ) = F ′ (X (k) ) − 3 F ′′ (X (k) )(e(k) ) + 3 F ′′ (X (k) ) (F ′ (X (k) )) F′′(X (k) )(e(k) ) 2 2 35 −1 1 2 3 2 − F ′′ (X (k) ) (F′(X (k) )) F′′′(X (k) )(e(k) ) + F ′′′ (X (k) )(e(k) ) 9 9 −1 −1 1 1 2 3 − F ′′′ (X (k) )(e(k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + F ′′′ (X (k) )(e(k) ) (F′(X (k) )) F ′′′ (X (k) )(e(k) ) 9 27 2 −1 −1 1 1 ′′′ (k) 3 (k) (k) 2 + F ′′′ (X (k) )(e(k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + F (X ) ((F ′ (X (k) )) F ′′(X )(e ) ) 9 18 −1 −1 1 ′′′ (k) 2 3 F (X ) [(F ′ (X (k) )) F ′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] 54 −1 1 4 + F ′′′ (X (k) ) (F′(X (k) )) F ′′′ (X (k) )(e(k) ) 27 −1 −1 1 3 2 − F ′′′ (X (k) ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′ (X (k) )(e(k) ) ] 54 − + 2 −1 1 ′′′ (k) 3 F (X ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] 162 − −1 4 (iv) (k) 1 2 F (X ) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 81 2! 3 −1 1 3 4 4 + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )) + O (‖e(k) ‖ ) 3! 3 −1 1 Let O (‖e(k) ‖ ) = − 9 F′′(X (k) ) (𝐅 ′ (X (k) )) − F ′′′ (X (k) )(e(k) ) 3 −1 1 ′′′ (k) (k) 2 F (X )(e ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 9 + ⋯ + (− −1 4 (iv) (k) 1 2 F (X )) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 81 2! 3 −1 1 3 4 4 + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )) + O (‖e(k) ‖ ) 3! Then, ⟹ F′ ( X(k) +2𝐘 (𝐤) 3 2 2 2 1 −1 ) = F ′ (X (k) ) − 3 F ′′ (X (k) )(e(k) ) + 3 F ′′ (X (k) ) (F ′ (X (k) )) 3 + 9 F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) F ′′ (X (k) )(e(k) ) 2 (6.20) Now using equations (6.16) and (6.20) in equation (6.15), we have ′ [F (X ′ (k) = F (X X (k) + 2Y (k) X (k) + 2Y (k) (k+1) ′ (k) ′ ) + 3F ( )] e = [F (X ) + 3F ( )] (e(k) ) − 4F(X (k) ) 3 3 (k) ′ )(e (k) X (k) + 2Y (k) ) + 3F ( ) (e(k) ) − 4F(X (k) ) 3 ′ 36 = F ′ (X (k) )(e(k) ) −1 2 1 2 +3 [F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 3 3 2 2 3 + F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )] (e(k) ) 9 1 1 2 3 4 −4 [F ′ (X (k) )(e(k) ) − F ′′ (X (k) )(e(k) ) + F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )] 2! 3! = F ′ (X (k) )(e(k) ) + 3F ′ (X (k) )(e(k) ) − 2F ′′ (X (k) )(e(k) ) 2 2 3 4 + F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) 3 2 2 3 4 −4F′ (X(k) )(e(k) ) + 2F′′ (X(k) )(e(k) ) − F′′′ (X(k) )(e(k) ) + O (‖e(k) ‖ ) 3 +F ′′ (X (k) ) (F ′ (X (k) )) −1 F ′′(X (k) )(e(k) )3 Finally, after cancellation of like terms, we obtain the following. ′ [F (X (k) X (k) + 2Y (k) ) + 3F ( )] e(k+1) 3 ′ −1 = [F ′′ (X (k) ) (F ′ (X (k) )) 3 4 F′′(X (k) )] (e(k) ) + O (‖e(k) ‖ ) Therefore, this is the proof of algorithm (6.2). From this proof of error equation, one can see that Algorithm (6.2) has a cubic order of convergence. 6.2.2. Proof of cubic order of convergence for algorithm 6.4 Theorem 6.2 (Noor and Waseem, 2008): Let F: D⊆ Rn → Rn , be r- times Fréchet differentiable function in an open and convex set D⊆ Rn containing the root 𝜶 of F(X) = 0. Suppose that F ′ (X (k) ) is continuous and nonsingular in D. If the initial approximate value X (𝟎) is sufficiently close to the exact root 𝜶, then the iterative method defined by Algorithm 6.4 converges to 𝜶 with cubic order of convergence and satisfies the error equation 2X (k) + Y (k) [3F ′ ( ) + F ′ (Y (k) )] e(k+1) 3 −1 = [F ′′ (X (k) ) (F ′ (X (k) )) 3 4 F ′′ (X (k) ] (e(k) ) + O (‖e(k) ‖ ) Proof: The proof is adopted from Noor and Waseem, 2008. From Algorithm (6.4) the method is given by the formula: (6.21) 37 2X(k) +𝐘 (𝐤) X (k+1) = X (k) − 4 [3F ′ ( 3 ) + F ′ (Y (k) )] −1 wher𝑒 Y (k) = X (k) − (F ′ (X (k) )) −1 F(X (k) ) (6.22) F ′ (X (k) ). Here we use Y (k) for X (k+1) to show the difference between X (k+1) of predictor step from X (k+1) of the corrector step. Now from equation (6.22), we have 2X(k) +𝐘 (𝐤) X (k+1) − X (k) = −4 [3F ′ ( 3 ) + F ′ (Y (k) )] −1 F(X (k) ) (6.23) But from equation (6.13), we know that X (k) = e(k) + 𝜶 ⟹ X (k) − 𝜶 = e(k) −1 ⟹X (k+1) − (e (k) 2X (k) + Y (k) + 𝜶) = −4 [3F ( ) + F ′ (Y (k) )] 3 ′ F(X (k) ) −1 ⟹ (X (k+1) − 𝜶) − e (k) 2X (k) + Y (k) = −4 [3F ( ) + F ′ (Y (k) )] 3 ′ 2X(k) +𝐘 (𝐤) ⟹ e(k+1) − e(k) = −4 [3F ′ ( 3 Pre-multiplying both sides by[3F ′ ( ) + F ′ (Y (k) )] 2X(k) +𝐘 (𝐤) 3 −1 F(X (k) ) F(X (k) ) (6.24) ) + F ′ (Y (k) )], we obtain 2X (k) + Y (k) [3F ( ) + F ′ (Y (k) ] (e(k+1) − e(k) ) = −4F(X (k) ) 3 ′ 2X(k) +Y(k) ⟹ [3F ′ ( 3 2X(k) +𝐘 (𝐤) ) + F ′ (Y (k) )] e(k+1) = [3F ′ ( 2X(k) +Y(k) Now applying a Taylor series on F ′ ( F′ ( 3 3 ) + F ′ (Y (k) )] e(k) − 4F(X (k) ) ) at the point X (k) , we have: 2X (k) + Y (k) 2X (k) + Y (k) ) = F ′ (X (k) ) + F ′′ (X (k) ) ( − X (k) ) 3 3 2 3 1 ′′′ (k) 2X (k) + Y (k) 1 (iv) (k) 2X (k) + Y (k) (k) + F (X ) ( − X ) + F (X ) ( − X (k) ) 2! 3 3! 3 4 2X (k) + Y (k) +O (‖ − X (k) ‖ ) 3 2 ′ = F (X (k) ′′ ) + F (X (k) −X (k) + Y (k) 1 −X (k) + Y (k) )( ) + F ′′′ (X (k) ) ( ) 3 2! 3 (6.25) 38 3 4 1 −X (k) + Y (k) −X (k) + Y (k) + F (iv) (X (k) ) ( ) + O (‖ ‖ ) 3! 3 3 2X (k) + Y (k) 1 1 2 F ( ) = F ′ (X (k) ) + F ′′ (X (k) )(Y (k) − X (k) ) + F ′′′ (X (k) )(Y (k) − X (k) ) 3 3 18 ′ + 1 (iv) (k) 3 4 F (X )(Y (k) − X (k) ) + O (‖e(k) ‖ ) 162 We can write the above equation as follows by replacing Y (k) by X (k+1) F′ ( 2X (k) + Y (k) 1 1 ′′′ (k) 2 ) = F ′ (X (k) ) + F ′′ (X (k) )(X (k+1) − X (k) ) + F (X )(X (k+1) − X (k) ) 3 3 18 3 1 4 + 162 F (iv) (X (k) )(X (k+1) − X (k) ) + O (‖e(k) ‖ ) (6.26) Now substituting equation (6.18) in equation (6.26), we have −1 2X (k) + Y (k) 1 F ( ) = F ′ (X (k) ) + F ′′ (X (k) ) (− (F ′ (X (k) )) F ′ (X (k) )) 3 3 ′ 2 −1 −1 1 1 (iv) (k) ( ) ( ) + F ′′′ (X (k) ) (− (F ′ (X (k) )) F ′ (X (k) )) + F (X ) (− (F′ (X k )) F′ (X k )) 18 162 3 4 +O (‖e(k) ‖ ) 2X(k) +Y(k) F′ ( 1 3 ′′′ + 18 F (X 1 ( ) −1 ) = F ′ (X (k) ) − 3 F ′′ (X (k) ) ((F′ (X k )) (k) ′ ) ((F (X 4 +O (‖e(k) ‖ ) (k) )) −1 F′ (X(k) )) 2 ′ F (X (k) 1 )) − 162 F (iv) (X (k) ′ ) ((F (X (k) −1 )) 3 ′ F (X (k) )) (6.27) Again substituting equation (6.17a) in equation (6.27), one can obtain the following result: F′ ( 2X (k) + Y (k) )= 3 1 1 ′ (k) −1 ′′ (k) 1 ′ (k) −1 ′′′ (k) 2 3 4 F ′ (X (k) ) − F ′′ (X (k) ) ((e(k) ) − (F (X )) F (X ) (e(k) ) + (F (X )) F (X ) (e(k) ) + O (‖e(k) ‖ )) 3 2! 3! −1 −1 1 1 1 2 3 4 + F ′′′ (X (k) ) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )) 18 2! 3! 2 39 −1 −1 1 (iv) (k) 1 1 2 3 F (X ) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 162 2! 3! − 4 3 4 + O (‖e(k) ‖ )) + O (‖e(k) ‖ ) −1 2X (k) + Y (k) 1 1 2 F ( ) = F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 3 3 6 ′ −1 1 ′′ (k) 3 4 ( ) ( ) F (X ) (F′ (X k )) F′′′ (X k )(e(k) ) + O (‖e(k) ‖ ) 18 − 2 1 −1 1 + 18 F ′′′ (X (k) ) [(e(k) ) − 2! (e(k) ) (F ′ (X (k) )) −1 1 (e(k) ) (F ′ (X (k) )) 3! −1 1 ((F ′ (X (k) )) 4 1 [(F ′ (X (k) )) 12 (F ′ (X (k) )) 3! 4 −1 1 (F ′ (X (k) )) 2! 3 F ′′ (X (k) )(e(k) ) + 2 2 2 F ′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) −1 3 F ′′′ (X (k) )(e(k) ) ] + 4 F ′′′ (X (k) )(e(k) ) − 1 −1 1 −1 [(F ′ (X (k) )) 12 [(F ′ (X (k) )) 36 3 F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) + F ′′ (X (k) )(e(k) ) ) − −1 −1 1 2 F ′′ (X (k) )(e(k) ) + −1 3 F ′′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) 3 2 2 F ′′ (X (k) )(e(k) ) ] + 4 F ′′′ (X (k) )(e(k) ) ] + O (‖e(k) ‖ )] +. . . . −1 2X (k) + Y (k) 1 1 2 F ( ) = F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F′′(X (k) )(e(k) ) 3 3 6 ′ − −1 1 ′′ (k) 1 3 2 ( ) ( ) F (X ) (F′ (X k )) F′′′ (X k )(e(k) ) + F ′′′ (X (k) )(e(k) ) 18 18 − −1 −1 1 ′′′ (k) (k) 1 ′′′ (k) (k) 2 3 F (X )(e ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + F (X )(e ) (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 36 108 2 −1 −1 1 ′′′ (k) 1 ′′′ (k) (k) (k) (k) 3 (k) (k) (k) 2 ′ ′′ ′ ′′ + F (X ) (F (X )) F (X )(e ) + F (X ) [(F (X )) F (X )(e ) ] 36 72 −1 −1 1 ′′′ (k) 2 3 − F (X ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] 216 −1 1 ′′′ (k) 4 + F (X ) (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 108 −1 −1 1 ′′′ (k) 3 2 − F (X ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] 216 40 + 2 −1 1 ′′′ (k) 3 4 F (X ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] + O (‖e(k) ‖ ) + ⋯ 648 3 Suppose that O (‖e(k) ‖ ) = − 18 F ′′ (X (k) ) (F′ (X k )) 1 ( ) −1 F′′′ (X k )(e(k) ) ( ) 3 −1 1 ′′′ (k) (k) 2 F (X )(e ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 36 −1 1 ′′′ (k) (k) 3 + F (X )(e ) (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 108 − +⋯+ 2 −1 1 ′′′ (k) 3 4 F (X ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] + O (‖e(k) ‖ ) + ⋯ 648 Then the above equation becomes −1 2X (k) + Y (k) 1 1 2 F ( ) = F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 3 3 6 ′ + 1 18 2 3 F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) (6.28) Now using equation (6.16) and (6.28) in (6.25), we have [3F ′ ( 2X (k) + Y (k) 2X (k) + Y (k) ) + F ′ (Y (k) )] e(k+1) = [3F ′ ( ) + F ′ (Y (k) )] e(k) − 4F(X (k) ) 3 3 2X(k) +Y(k) ⟹[3F ′ ( 3 2X(k) +Y(k) [3F ′ ( 3 1 2X(k) +Y(k) ) + F ′ (Y (k) )] e(k+1) = 3F ′ ( 3 ) (e(k) ) + F ′ (Y (k) )(e(k) ) − 4F(X (k) ) 1 ) + F ′ (Y (k) )] e(k+1) = 3 [F ′ (X (k) ) − 3 F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) 6 −1 2 1 2 3 1 F ′′′ (X (k) )(e(k) ) ] + O (‖e(k) ‖ ) F ′′ (X (k) )(e(k) ) + 18 F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )] (e(k) ) + F ′ (Y (k) )(e(k) ) 1 2 −4 [F ′(X (k) )(e(k) ) − 2! F ′′ (X (k) )(e(k) ) + 3! 3 4 (6.29) But by applying a Taylor series in equation (6.2) for F′(Y(k) ) at a point X (k) , one can obtain the following. F ′ (Y (k) ) = F′(X (k) ) + F ′ ′(X (k) )(Y (k) − X (k) ) + 1 ′′ (k) 2 3 F ′(X )(Y (k) − X (k) ) + O (‖Y (k) − X (k) ‖ ) 2! −1 Since Y (k) = X (k+1) and X (k+1) − X (k) = − (F ′ (X (k) )) −1 F ′ (Y (k) ) = F′(X (k) ) + F ′ ′(X (k) ) (− (F ′ (X (k) )) F ′ (X (k) ) then, F ′ (X (k) )) 41 2 −1 1 3 + F ′′ ′(X (k) ) (− (F ′ (X (k) )) F ′ (X (k) )) + O (‖e(k) ‖ ) 2! −1 𝐹 ′ (Y (k) ) = F ′ (X (k) ) − F ′ ′(X (k) ) ((F ′ (X (k) )) 1 ′′ + 2! F ′(X (k) ′ ) ((F (X (k) −1 )) F ′ (X (k) )) 2 ′ F (X (k) 3 )) + O (‖e(k) ‖ ) (6.30) From equation (6.17a), we have (F ′ (X (k) )) −1 1 F(X (k) ) = e(k) − −1 + 3! (F ′ (X (k) )) 1 ′ (k) −1 ′′ (k) (k) 2 (F (X )) F (X )(e ) 2! F ′′′ (X (k) )(e(k) ) 3 4 + O (‖e(k) ‖ ) (*) Now substitute (*) in equation (6.30), the result becomes F ′ (Y (k) ) = F ′ (X (k) ) − F′′(X (k) ) ((e(k) ) − + + 1 ′ (k) −1 ′′ (k) (k) 2 (F (X )) F (X )(e ) 2! 1 ′ (k) −1 ′′′ (k) (k) 3 4 (F (X )) 𝐅 (X )(e ) + O (‖e(k) ‖ )) 3! −1 −1 1 ′′ (k) 1 1 2 3 F ′(X ) ((e(k) ) − (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 2! 2! 3! + O (‖e (k) 4 2 F ′ (Y (k) ) = F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + − 3 ‖ )) + O (‖e(k) ‖ ) −1 1 ′′ (k) 2 F (X ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 2! −1 1 4 3 F′′(X (k) ) (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) 3! 42 + 1 2! 2 F′′′ (X(k) ) [(e(k) ) − + + − + − + 1 3! 1 2! 1 3! 1 2! (e(k) ) (F′ (X(k) )) (e(k) ) (F′ (X(k) )) (F′ (X(k) )) 12 1 1 [(F′ (X(k) )) (F′ (X(k) )) 12 1 36 −1 [(F′ (X(k) )) ′ (k) [(F (X )) 3 3 −1 1 4 4 ((F′ (X(k) )) 2 F′′ (X(k) )(e(k) ) ] [(F′ (X(k) )) F′′′ (X(k) )(e(k) ) −1 2 F′′′ (X(k) )(e(k) ) + O (‖e(k) ‖ ) −1 3 ′′′ (k) F (X −1 2 2 F′′(X(k) )(e(k) ) ) 3 F′′′ (X(k) )(e(k) ) ] 4 F′′′ (X(k) )(e(k) ) ] [(F′ (X(k) )) F ′ (Y (k) ) = F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + − F′′ (X(k) )(e(k) ) F′′(X(k) )(e(k) ) + −1 −1 −1 −1 (k) 3 2 −1 2 F′′ (X(k) )(e(k) ) ] 3 4 )(e ) ] + O (‖e(k) ‖ )] + O (‖e(k) ‖ ) −1 1 ′′ (k) 2 ( ) ( ) F (X ) (F′ (X k )) F′′ (X k )(e(k) ) 2! −1 1 3 4 F′′(X(k) ) (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) 3! −1 1 2 2 F′′′(X(k) )(e(k) ) − F′′′(X(k) )(e(k) ) (F′ (X(k) )) F′′ (X(k) )(e(k) ) 2! 4 −1 1 3 4 + F′′′(X(k) )(e(k) ) (F′ (X(k) )) F′′′ (X(k) )(e(k) ) + O (‖e(k) ‖ ) 12 + 1 2 −1 −1 1 1 3 2 + F′′′(X (k) ) (F ′ (X (k) )) F′′(X (k) )(e(k) ) + F′′′(X (k) ) ((F ′ (X (k) )) F′′(X (k) )(e(k) ) ) 4 8 −1 −1 1 2 3 − F′′′(X (k) ) [(F ′ (X (k) )) F ′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] 24 −1 1 4 + F′′′(X (k) ) (F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) 12 −1 −1 1 3 2 − F′′′(X (k) ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] [(F ′ (X (k) )) F ′′ (X (k) )(e(k) ) ] 24 + 2 −1 1 3 4 F′′′(X (k) ) [(F ′ (X (k) )) F ′′′ (X (k) )(e(k) ) ] + O (‖e(k) ‖ ) 72 3 1 ( ) −1 Assume that, O (‖e(k) ‖ ) = − 3! F′′(X (k) ) (F′ (X k )) F′′′ (X(k) )(e(k) ) 3 43 −1 1 2 − F′′′(X (k) )(e(k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 4 +⋯+ 1 72 (k) F′′′(X ′ (k) ) [(F (X )) −1 ′′′ (k) F (X (k) 3 2 4 )(e ) ] + O (‖e(k) ‖ ) 1 ( ) −1 Then, F ′ (Y (k) ) = F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + 2! F ′′ (X (k) ) (F′ (X k )) 2 1 ( ) F′′ (X k )(e(k) ) 3 + F′′′ (X(k) )(e(k) ) + O (‖e(k) ‖ ) (6.31) 2! Now, by replacing equation (6.31) in equation (6.29) in the place of F ′ (Y(k) ), we have 2X (k) + Y (k) [3F ( ) + F ′ (Y (k) )] e(k+1) 3 ′ −1 1 1 2 = 3 [F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 3 6 1 2 3 + F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )] (e(k) ) 18 −1 1 2 + [F ′ (X (k) ) − F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 2! 1 2 3 + F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )] (e(k) ) 2! 1 1 2 3 4 −4 [F ′ (X (k) )(e(k) ) − F ′′ (X (k) )(e(k) ) + F ′′′ (X (k) )(e(k) ) + O (‖e(k) ‖ )] 2! 3! 2X (k) + Y (k) ⟹ [3F ( ) + F ′ (Y (k) )] e(k+1) 3 ′ −1 1 2 2 = 3F ′ (X (k) )(e(k) ) − F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) 2 1 3 2 + F ′′′ (X (k) )(e(k) ) + F ′ (X (k) )(e(k) ) − F ′′ (X (k) )(e(k) ) 6 −1 1 + 2! F ′′ (X (k) ) (F ′ (X (k) )) 2 +2F ′′ (X (k) )(e(k) ) − 2 3 1 3 F ′′ (X (k) )(e(k) ) + 2! F ′′′ (X (k) )(e(k) ) − 4F ′ (X (k) )(e(k) ) 2 ′′′ (k) (k) 3 4 F (X )(e ) + O (‖e(k) ‖ ) 3 44 2X (k) + Y (k) ⟹ [3F ′ ( ) + F ′ (Y (k) )] e(k+1) = (4F ′ (X (k) )(e(k) ) − 4F ′ (X (k) )(e(k) )) 3 +( −1 −1 1 ′′ (k) 1 3 3 F (X ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) ) 2! 2! 2 2 + (−2F ′′ (X (k) )(e(k) ) + 2F ′′ (X (k) )(e(k) ) ) 1 1 2 3 3 3 4 + [ F ′′′ (X (k) )(e(k) ) + F ′′′ (X (k) )(e(k) ) − F ′′′ (X (k) )(e(k) ) ] + O (‖e(k) ‖ ) 6 2 3 After cancellation of like terms, we obtain −1 2X (k) + Y (k) 3 4 [3F ′ ( ) + F ′ (Y (k) )] e(k+1) = F ′′ (X (k) ) (F ′ (X (k) )) F ′′ (X (k) )(e(k) ) + O (‖e(k) ‖ ) 3 Therefore, this is the proof for error equation of Algorithm (6.4). From this proof one can understand that Algorithm (6.4) has cubic order of convergence. 45 7. NUMERICAL RESULTS OF THE FOUR METHODS In this chapter, some numerical examples are given to show the performance of the four methods (Newton, Broyden, PCM-1 and PCM-2 methods) and the number of the iteration results of the one method is compared to the other among them. For the sake of comparison, the given examples are also solved by the four methods. Each value of the iteration of the methods is given in a table and then the comparison is made on the other table for all methods in common. Now all the examples given below are tested with a precision 𝜀 = 10−4 and the stopping criteria of the iteration process for the given examples is ‖X (k) − X (k−1) ‖2 < 𝜀. To show how the methods perform, some of the examples are solved manually and the rest are solved by using MATLAB codes given to the methods to ease the work. Example 7.1 Consider the following system of nonlinear equations (Hafiz and Bahgat, (2012)) and find the approximate solution X (k+1) of the system if the initial value x (0) = [0.1 0.1 −0.1]𝑇 . 𝑥12 F(X) = [ 3𝑥1 − cos(𝑥2 𝑥3 ) − 0.5 0 − 81(𝑥2 + 0.1)2 + sin(𝑥3 ) + 1.06 = [0] 10𝜋 − 3 −𝑥1 𝑥2 0 e + 20𝑥3 + ] 3 Solution by using Newton’s method The Newton’s method is given by: −1 X (k+1) = X (k) − (F ′ (X (k) )) F(X (k) ) Now evaluating F(X) and F′(X) at x (0) , we have 3(0.1) − cos(0.1 × −0.1) − 0.5 −1.2000 (0.1) − 81(0.1 + 0.1)2 + sin(−0.1) + 1.06 F(x (0) ) = = [−2.2698] 10𝜋 − 3 −(0.1×0.1) 8.4620 e + 20(−0.1) + [ ] 3 2 46 F ′ (X) 3 2𝑥1 =[ −𝑥2 e−𝑥1𝑥2 𝑥3 sin(𝑥2 𝑥3 ) −162(𝑥2 + 0.1) −𝑥1 e−𝑥1 𝑥2 𝑥2 sin(𝑥2 𝑥3 ) cos(𝑥3 ) ] 20 3 −0.1 × sin(0.1 × −0.1) 2 × 0.1 −162(0.1 + 0.1) F (x ) = [ −(0.1×0.1) −0.1 × e −0.1 × e−(0.1×0.1) 3.0000 0.0010 −0.0010 = [ 0.2000 −32.4000 0.9950 ] −0.0990 −0.0990 20.0000 ′ (0) 0.1 × sin(0.1 × −0.1) cos(−0.1) ] 20 Then, −1 X (1) = X (0) − (F ′ (X (0) )) X (1) F(X (0) ) 0.1 3.0000 = [ 0.1 ] − ([ 0.2000 −0.1 −0.0990 0.0010 −32.4000 −0.0990 −0.0010 −1 −1.2000 0.4999 ]) × [ ] = [ 0.9950 −2.2698 0.0195 ] 20.0000 8.4620 −0.5215 Taking the same procedure for the next iterations, their values are given in the following table. Table 1: Iteration result of Newton’s method for Example 7.1 k (𝒌) (𝒌) ‖𝐱(𝐤) − 𝐱(𝐤−𝟏) ‖𝟐 (𝒌) 0 𝒙𝟏 0.1000 𝒙𝟐 0.1000 𝒙𝟑 -0.1000 __ 1 0.4999 0.0195 -0.5215 0.5866 2 0.5000 0.0016 -0.5236 0.0180 3 0.5000 0.0000 -0.5236 0.0016 4 0.5000 0.0000 -0.5236 0.0000 From the table 1 it is clear that the iteration process repeats the same result after the third iteration because the solution in this project is approximated using 4 decimal places after the decimal point. The norm when 3rd iteration subtracted from the 4th is zero so that this result is less than the given 𝜺. One can understand from this the solution of the given system has reached its converging point and so that the approximate solution for the system is[0.5000 0.0000 −0.5236]𝑡 . 47 Solution by using the predictor-corrector method-1 (PCM-1) (or Algorithm 6.3) Now by using the iterations of Newton’s method illustrated in the table-1 as a predictor step for respective iterations of the corrector step of PCM-1, it is possible to find the approximate solution of the system of nonlinear equations given in example 7.1 by PCM-1 (the method in Algorithm 6.3) as follows. The formula of this method to solve such problems is given by Predictor step: Y (k) = X(k+1) = X(k) − (F′ (X(k) )) −1 F(X(k) ) (Newton’s method) And, The corrector step: X (k+1) X (k) + 2 Y(k) (k) ′ (k) ′ = X − 4 × [F (X ) + 3F ( )] 3 −1 F(X (k) ), where, k = 0,1,2, … Now the preceding conditions before coming to the iteration process is finding F ′ (X) of the X(k) +2 𝐘(𝐤) given system and then evaluating F(X) and F ′ (X) at x (0) . Again evaluating 3F ′ ( at x (0) and Y (0) . That is: F ′ (X) = [ 3 2𝑥1 −𝑥2 e−𝑥1𝑥2 𝑥3 sin(𝑥2 𝑥3 ) −162(𝑥2 + 0.1) −𝑥1 e−𝑥1 𝑥2 𝑥2 sin(𝑥2 𝑥3 ) cos(𝑥3 ) ] 20 3(0.1) − cos(0.1 × −0.1) − 0.5 −1.2000 (0.1) − 81(0.1 + 0.1)2 + sin(−0.1) + 1.06 F(x (0) ) = = [−2.2698] 10𝜋 − 3 −(0.1×0.1) 8.4620 e + 20(−0.1) + [ ] 3 3 −0.1 × sin(0.1 × −0.1) 0.1 × sin(0.1 × −0.1) 2 × 0.1 −162(0.1 + 0.1) cos(−0.1) F ′ (x (0) ) = [ ] −0.1 × e−(0.1×0.1) −0.1 × e−(0.1×0.1) 20 3.0000 0.0010 −0.0010 = [ 0.2000 −32.4000 0.9950 ] −0.0990 −0.0990 20.0000 2 And 3 ) 48 X (k) + 2 Y 3F ( 3 (k) ′ X (0) + 2 Y ) = 3F ( 3 (0) ′ ) But Y (0) is the first iteration value of Newton’s method so that, 0.4999 Y (0) = X (1) = [ 0.0195 ] (From Newton’s iteration) −0.5215 So, 3F ′ ( X(k) +2 Y(k) 3 1.0998 [ 0.1390 ] −1.1430 3 = 3 × F′ ( = 3[ X(0) +2 Y(0) ) = 3F ′ ( )= 3× ) = 3 × F′(0.3666 , 0.0463, −0.3810) ) 3 2 × 0.3666 −0.0463 × 𝑒 −0.3666×0.0463 9 = [ 2.1996 −0.1366 3 0.1 0.4999 [ 0.1 ]+ 2×[ 0.0195 ] −0.5215 F ′ ( −0.1 3 0.0202 −71.1018 −1.0813 −0.381 × sin(0.0463 × −0.381) −162(0.0463 + 0.1) −0.3666 × 𝑒 −0.3666×0.0463 0.0463 × sin(0.0463 × −0.381) ] cos(−0.3810) 20 −0.0025 2.7849 ] 60.0000 Now the corrector step becomes, ⟹ X (1) X (1) X (1) X (0) + 2 Y(0) (0) ′ (0) ′ = X − 4 × [F (X ) + 3F ( )] 3 0.1 3 0.001 = [ 0.1 ] − 4 ([ 0.2 −32.4 −0.1 −0.099 −0.099 −1.2000 × [−2.2698] 8.4620 0.1 12 = [ 0.1 ] − [ 2.3996 −0.1 −0.2356 0.0202 −103.5018 −1.1803 −1 F(X (0) ) −0.001 9 0.995 ] + [ 2.1996 20 −0.1366 0.0202 −71.1018 −1.0813 −0.0025 −1 2.7849 ]) 60.0000 −4.8 −0.0035 −1 3.7799 ] × [−9.0792] 33.848 80 0.5000 X (1) = [ 0.0061 ] −0.5233 Following the same manner for the next iterations, the solution is given in table 2. 49 Table 2: The iteration result of PCM-1 for Example 7.1 (𝒌) K (𝒌) 𝒙𝟏 (𝒌) 𝒙𝟑 -0.1000 ‖𝐱 (𝐤) − 𝐱 (𝐤−𝟏) ‖𝟐 __ 0 0.1000 𝒙𝟐 0.1000 1 0.5000 0.0061 -0.5233 0.5899 2 0.5000 0.0000 -0.5236 0.0016 3 0.5000 0.0000 -0.5236 0.0000 From this table we can see that the iteration process repeated the same result after the 2nd iteration and the Euclidean norm of the difference of 2nd iteration from 3rd is 0. This implies that the iteration converged to its approximate solution so that the approximate solution for the SNLEs given in Example 7.1 is [0.5000 0.0000 −0.5236]𝑇 . Solution by using the predictor-corrector method-2 (PCM-2) (or Algorithm 6.5) In this case also we are going to find an approximate solution for the given in Example 7.1. This method also gives the same approximate solution like other methods listed above. As PCM-1 did, this method also has the predictor and corrector steps with the form as follows. −1 Predictor step: Y (k) = X (k) − (F ′ (X (k) )) F(X (k) ) In this case also the predictor step is the Newton’s method. So all the values of Y (k) are the iteration values of Newton’s method. Corrector step: X (k+1) =X (k) − 2X(k) +𝐘 (𝐤) 4 [3F ′ ( ) 3 ′ + F (Y (k) −1 )] F(X (k) ), 𝑘 = 0, 1,2, … To find the approximate solution using this method we have given the initial approximation value of the system x (0) = [0.1 0.1 −0.1]𝑇 and Y (0) = [0.4999 0.0195 −0.5215]T . The value of F(X (0) ) is already calculated for the above methods and has the value, F(X (0) −1.2000 ) = [−2.2698]. 8.4620 2X(k) +Y(k) Now evaluating 3F ′ ( 3 ) for k=0, we have: 50 3F ′ ( 0.1000 0.4999 2 × [ 0.1000 ] + [ 0.0195 ] −0.1000 −0.5215 3 2X (0) + Y (0) ) = 3 × F′ 3 ( 0.6999 [ 0.2195 ] −0.7215 3 = 3 × F′ ) ( ) = 3 × F ′ (0.2333, 0.0732, −0.2405 ) But, F ′ (X) =[ 3 2𝑥1 −𝑥2 e−𝑥1 𝑥2 3 × F ′ (0.2333, 𝑥3 sin(𝑥2 𝑥3 ) 𝑥2 sin(𝑥2 𝑥3 ) −162(𝑥2 + 0.1) cos(𝑥3 ) ] then, −𝑥1 𝑥2 −𝑥1 e 20 −0.2405 ) 0.0732, 3 = 3[ 2 × 0.2333 −0.0732 × e−0.2333×0.0732 9.0000 = [ 1.3998 −0.2160 0.0126 −84.1752 −0.6879 −0.2405 × sin(0.0732 × −0.2405) 0.0732 × sin(0.0732 × −0.2405) −162(0.0732 + 0.1) cos(−0.2405) ] −0.2333 × e−0.2333×0.0732 20 −0.0039 2.9136 ] 60 And, F ′ (Y (0) ) = F′(0.4999, 0.0195, −0.5215) 3 2 × 0.4999 =[ −0.0195 × e−0.4999×0.0195 ′ F (Y (0) 3 ) = [ 0.9998 −0.0193 −0.5215 × sin(0.0195 × −0.5215) 0.0195 × sin(0.0195 × −0.5215) −162(0.0195 + 0.1) cos(−0.5215) ] −0.4999×0.0195 −0.4999 × e 20 0.0053 −19.3590 −0.4951 −0.0002 0.8671 ] 20 Here again, 2X(0) +Y(0) 3F ′ ( 3 ) + F ′ (Y (0) ) 9.0000 = [ 1.3998 −0.2160 0.0126 −84.1752 −0.6879 12 = [ 2.3996 −0.2353 0.0179 −103.5342 −1.1830 −0.0039 3 2.9136 ] + [ 0.9998 60 −0.0193 −0.0041 3.7807 ] 80.0000 Now the first iteration of PCM-2 becomes, 0.0053 −19.3590 −0.4951 −0.0002 0.8671 ] 20 51 −1 X (1) 2X (0) + Y (0) (0) ′ = X − 4 [3F ( ) + F ′ (Y (0) )] 3 X (1) 0.1000 12 = [ 0.1000 ] − 4 × ([ 2.3996 −0.1000 −0.2353 F(X (0) ) 0.0179 −103.5342 −1.1830 −0.0041 −1 −1.2000 3.7807 ]) × [−2.2698] 80.0000 8.4620 0.5000 X (1) = [ 0.0061 ] −0.5233 Following the same manner as it is done in the 1st iteration, the solution is given in table 3. Table 3: The iteration result of PCM-2 for Example 7.1 (𝒌) K (𝒌) 𝒙𝟏 𝒙𝟐 (𝒌) 𝒙𝟑 ‖𝐱 (𝐤) − 𝐱 (𝐤−𝟏) ‖𝟐 0 0.1000 0.1000 -0.1000 ___ 1 0.5000 0.0061 -0.5233 0.5899 2 0.5000 0.0000 -0.5236 0.0061 3 0.5000 0.0000 -0.5236 0.0000 As it is shown in the table above, the iteration process for the approximate solution repeats the same iteration values after the 2nd iteration and the 2-norm for the difference of the 2nd iteration from the 3rd iteration is zero. From these things one can understand that the iteration process has reached the convergence point. Therefore, the approximate solution by using the PCM-2 for the system given in Example 7.1 0.5000 is X = [ 0.0000 ]. −0.5236 Approximate solution for Example 7.1 by Broyden’s method This iterative method is given by the formula: X (k+1) = X (k) − (B(k) ) −1 (F(X (k) )) 52 But B(𝑘) = B(𝑘−1) + 𝐹(X(k) )(𝛿 (𝑘−1) ) 𝑇 𝑇 (𝛿 (𝑘−1) ) (𝛿 (𝑘−1) ) , 𝑘 = 0, 1, 2, …, and B(𝑘) is an approximate updating matrix instead of Jacobian matrix which has nearly the same value with Jacobian matrix. The initial updating matrix B (0) is always the same with the value of the Jacobian matrix evaluated at X (0) . So having this concept the SNLEs given in Example 7.1 is solved as follows for the same initial value X (0) . −1.2000 Then, F(X (0) ) = [−2.2698] 8.4620 And, B (0) = F ′ (x (0) ) = 3 2 × 0.1 [ −0.1 × e−(0.1×0.1) 3.0000 = [ 0.2000 −0.0990 −0.1 × sin(0.1 × −0.1) 0.1 × sin(0.1 × −0.1) −162(0.1 + 0.1) cos(−0.1) ] −(0.1×0.1) −0.1 × e 20 0.0010 −32.4000 −0.0990 −0.0010 0.9950 ] 20.0000 From equation (4.19) we have B(k) 𝛿 (𝑘) = −𝐹(X (k) ), where 𝛿 (𝑘) = X (k+1) − X (k) . ⟹B(0) 𝛿 (0) = −𝐹(X (0) ) 3.0000 [ 0.2000 −0.0990 𝛿 (0) 0.0010 −32.4000 −0.0990 3.0000 = ([ 0.2000 −0.0990 −0.0010 −1.2000 (0) 0.9950 ] 𝛿 = − [−2.2698] 20.0000 8.4620 0.0010 −32.4000 −0.0990 −0.0010 −1 1.2000 0.3999 0.9950 ]) × [ 2.2698 ] = [−0.0805] 20.0000 −8.4620 −0.4215 But, 𝛿 (0) = X (1) − X (0) ⟹X (1) =X (0) +𝛿 (0) 0.1000 0.3999 0.4999 = [ 0.1000 ] + [−0.0805] = [ 0.0195 ] −0.1000 −0.4215 −0.5215 Continuing the iteration in the above manner and using B(𝑘) = B(𝑘−1) + k= 1, 2, 3, …, the solution at each iteration is given in the following table. 𝐹(X(k) )(𝛿 (𝑘−1) ) 𝑇 𝑇 (𝛿 (𝑘−1) ) (𝛿 (𝑘−1) ) for 53 Table 4: The summary of the iteration results for Example 7.1 by Broyden’s method (𝒌) k ‖𝐱(𝐤) − 𝐱(𝐤−𝟏) ‖𝟐 (𝒌) (𝒌) 𝒙𝟏 𝒙𝟑 𝒙𝟐 0 0.1000 0.1000 -0.1000 ___ 1 0.4999 0.0195 -0.5215 0.5866 2 0.5000 0.0088 -0.5232 0.0108 3 0.5000 0.0008 -0.5234 0.0080 4 0.5000 0.0000 -0.5236 0.00082462 5 0.5000 0.0000 -0.5236 0.0000 Thus from the table above, the norm is equal to zero at the 5th iteration. This implies that the SNLEs given in Example 7.1 has converged to its approximate solution and so that the system 0.5000 has the approximate solution X =[ ]. 0 −0.5236 Comparison of the results of the four methods listed above The following table gives the comparison of the methods how fast they converge to approximate solution based the number of iterations they perform. Table 5: Comparison of BM, NM, PCM-1 and PCM-2 for Example 7.1 F( X) NI Broyden method(BM) Newton method(NM) 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 1 2 0.4999 0.5000 0.0195 0.0088 -0.5215 -0.5232 0.4999 0.0195 0.0016 -0.5215 0.5000 0.0061 -0.5233 0.5000 0.0061 -0.5233 -0.5236 0.5000 0.0000 -0.5236 0.5000 0.0000 -0.5236 3 4 5 0.5000 0.5000 0.5000 0.0008 0.0000 0.0000 -0.5234 -0.5236 -0.5236 0.5000 0.0000 0.0000 -0.5236 0.5000 0.0000 -0.5236 0.5000 0.0000 -0.5236 0.5000 0.5000 PCM-1(Alg.6.3) PCM-2(Alg.6.5) 𝑥3 -0.5236 From the above table, it’s very important to see that the same given SNLEs has different number of iterations for different methods. As it illustrated above Broyden method takes 5 iterations for the system to converge and the Newton’s method takes 4 iterations for the same system. But the PCM-1 and PCM-2 methods take only three iterations until the system comes to converge to the approximate solution. 54 Therefore, based on the number of iterations that the four methods perform, it’s possible to say that the PCM-1 and PCM-2 methods are better than that of the other two methods compared with them. Example 7.2 Solve the following system of nonlinear equations which has the exact solution [1 1]𝑇 using initial approximation X (0) = [0.5 0.5]T (Hafiz and Mohammed, (2012)). The system is given as, F2 (X) = ( 𝑥12 − 10𝑥1 + 𝑥22 + 8 0 )=[ ] 0 𝑥1 𝑥22 + 𝑥1 − 10𝑥2 + 8 Solution by Newton’s method Just as it is done in Example 7.1, we follow the same manner to solve this system and its solution for the iterations is given in the following table. The above iteration results are shortly summarized in the table below. Table 6: The summary of iteration results of Newton’s method for Example 7.2 (𝒌) (𝒌) ‖𝐱(𝐤) − 𝐱(𝐤−𝟏) ‖𝟐 k 𝒙𝟏 𝒙𝟐 0 0.5000 0.5000 ___ 1 0.9377 0.9392 0.6201 2 0.9987 0.9984 0.0850 3 1.0000 1.0000 0.0021 4 1.0000 1.0000 0.0000 From the table above the norm is zero at the 4th iteration and the sequence repeats the same results after the 3rd iteration. This implies that the system F2 (X) has reached the point of convergence. Therefore, the system has the exact solution [1 1]𝑇 . 55 Solution by PCM-1 Table 7: The summary of iteration results for PCM-1 in Example 7.2 (𝒌) (𝒌) k ‖𝐱 (𝐤) − 𝐱 (𝐤−𝟏) ‖ 𝒙 𝒙 𝟏 𝟐 0 0.5000 0.5000 ____ 1 0.9911 0.9900 0.6937 2 1.0000 1.0000 0.0134 3 1.0000 1.0000 0.0000 𝟐 The norm at the 3rd iteration is equal to zero. This implies that the sequence is converged to the solution and therefore, the system F2 (X) has the exact solution X = [1 1]T . The solution of F2 by PCM-2 Table 8: The iteration result of PCM-2 for the system given in Example 7.2 (𝒌) (𝒌) ‖𝐱 (𝐤) − 𝐱 (𝐤−𝟏) ‖𝟐 k 𝒙𝟏 𝒙𝟐 0 0.5000 0.5000 ___ 1 0.9841 0.9483 0.6598 2 1.0000 1.0000 0.0541 3 1.0000 1.0000 0.0000 From the table given above the norm being zero at the 3rd iteration implies that the sequence of the iteration converged to the solution. So the value X = [1 1]𝑇 is the exact solution of the system given in Example 7.2. Solution by Broyden’s method Table 9: The iteration result of the Broyden’s method for the system given in Example 7.2 (𝒌) (𝒌) ‖𝐱(𝐤) − 𝐱(𝐤−𝟏) ‖𝟐 K 𝒙𝟏 0 0.5000 0.5000 ___ 1 0.9377 0.9392 0.6201 2 0.9912 0.9901 0.0738 3 0.9998 0.9997 0.0129 4 1.0000 1.0000 0.000036056 5 1.0000 1.0000 0.0000 𝒙𝟐 56 The norm at the 5th iteration is zero. This indicates that the system has converged to its solution. Therefore, the given system of nonlinear equations has the exact solutionX = [1 1]𝑇 . The following table gives the comparison of the methods how fast they converge to solution based on the number of iterations they perform for the given example 7.2. Table 10: Comparison of BM, NM, PCM-1 and PCM-2 for Example 7.2 𝐅𝟐 (X) Broyden method(BM) Newton method(NM) PCM-1(Alg.6.3) 𝑥2 PCM-2(Alg.6.5) NI 𝑥1 𝑥2 𝑥1 𝑥2 𝑥1 𝑥1 𝑥2 1 0.9377 0.9392 0.9377 0.9392 0.9911 0.9900 0.9841 0.9483 2 0.9912 0.9901 0.9987 0.9984 1.0000 1.0000 1.0000 1.0000 3 0.9998 0.9997 1.0000 1.0000 1.0000 1.0000 1.0000 1.0000 4 1.0000 1.0000 1.0000 1.0000 5 1.0000 1.0000 From the above table, it’s very important to see that the same given SNLEs has different number of iterations for different methods. As it illustrated above Broyden method takes 5 iterations for the system to converge and the Newton’s method takes 4 iterations for the same system. But the PCM-1 and PCM-2 methods take only three iterations until the system comes to converge to the exact solution of the system given in Example 7.2. Therefore, based on the number of iterations that the four methods perform, it’s nice to suggest that the PCM-1 and PCM-2 methods are better than the other two methods compared with them. Here below Examples 7.3 to 7.6 are given to show more on the performance of the four methods. These problems are solved using different MATLAB codes given for the four methods and these codes are written in appendices A, B, C, …, H in chapter-10. Example 7.3 [Darvishi and Barati, (2007)]: Solve the following system of nonlinear equations at initial approximate value(0.5 0.5 0.5)𝑇 . 57 𝑥12 + 𝑥22 + 𝑥32 − 1 = 0 2𝑥12 + 𝑥22 − 4𝑥3 = 0 3𝑥12 − 4𝑥22 + 𝑥32 = 0 Solution: The solution of this system is given in table-11 for all methods together. The system is solved using MATLAB codes that are given in Appendices A, B, C and D. Table 11: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.3 F(X) Broyden method(BM) Newton method(NM) PCM-1(Alg.6.3) PCM-2(Alg.6.5) NI 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 1 0.7500 0.6500 0.3500 0.7500 0.6500 0.3500 0.6906 0.6263 0.3429 0.7171 0.6374 0.3465 2 0.6885 0.6275 0.3444 0.7001 0.6289 0.3426 0.6983 0.6285 0.3426 0.6983 0.6285 0.3426 3 0.6985 0.6280 0.3428 0.6983 0.6285 0.3426 0.6983 0.6285 0.3426 0.6983 0.6285 0.3426 4 0.6982 0.6286 0.3426 0.6983 0.6285 0.3426 5 0.6983 0.6285 0.3426 6 0.6983 0.6285 0.3426 From the above table, we can see that the given system has different number of iterations when solved by using different methods. As it is illustrated above, Broyden’s method took 6 iterations for the system to converge and the Newton’s method took 4 iterations. But both PCM-1 and PCM-2 methods took only three iterations until the system comes to converge to the approximate solution and the approximate solution of the given system is 0.6983 X = [0.6285] 0.3426 Therefore, the PCM-1 and PCM-2 methods took less number of iterations than others. This implies both PCM-1 and PCM-2 methods are effective and better than Newton and Broyden’s methods. Example 7.4[Darvishi and Barati, (2004)]: Solve the following system of nonlinear equations at initial approximate value(2.5 0.5 1.5)𝑇 . 𝑥12 + 𝑥22 + 𝑥32 − 9 = 0 𝑥1 𝑥2 𝑥3 − 1 = 0 𝑥1 + 𝑥2 − 𝑥32 = 0 58 Solution: Using the same way of solving as we have done above, the solution for this case is also solved using MATLAB codes and the only iteration results for all methods are given in table-12 below. This system of nonlinear equations has been solved using the same codes in appendices A, B, C and D. The only changes made for this example is the use of different SNLEs with different initial approximation value. Table 12: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.4 F(X) Broyden method(BM) Newton method(NM) PCM-1(Alg.6.3) PCM-2(Alg.6.5) NI 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 𝑥1 𝑥2 𝑥3 1 2.5129 0.2114 1.6581 2.5129 0.2114 1.6581 2.4892 0.2459 1.6536 2.5003 0.2296 1.6559 2 2.4888 0.2461 1.6541 2.4917 0.2424 1.6535 2.4914 0.2427 1.6535 2.4914 0.2427 1.6535 3 2.4914 0.2429 1.6535 2.4914 0.2427 1.6535 2.4914 0.2427 1.6535 2.4914 0.2427 1.6535 4 2.4914 0.2427 1.6535 2.4914 0.2427 1.6535 5 2.4914 0.2427 1.6535 The results of the iteration process show from this table that the required approximate solution for the system given in Example 7.4 is[2.4914 0.2427 1.6535]T . Another thing what we can understand depending up on the number of iterations from the table-12 is also that there are differences in using the given methods. The methods with less number of iterations are better than that of methods took large number of iterations. In this case also the PCM-1 and PCM-2 are relatively better than Newton and Broyden’s methods. They took fewer numbers of iterations until they reach the point of convergence. Example 7.5 [Hafiz and Bahgat, (2012)] Consider the following system of nonlinear functions and find the solution of the system if the initial approximate solution 𝑥 (0) = [10 6 − 5]𝑇 . 𝑓1 (𝑥, 𝑦, 𝑧) = 15𝑥 + 𝑦 2 − 4𝑧 − 13 = 0 F(X) = {𝑓2 (𝑥, 𝑦, 𝑧) = 𝑥 2 + 10𝑦 − 𝑒 −𝑧 − 11 = 0 𝑓3 (𝑥, 𝑦, 𝑧) = 𝑦 3 − 25𝑧 + 22 = 0 Solution: The following table gives the iteration results of this problem for all methods given above and these results are found by using MATLAB codes in appendices A, B, C and D in chapter-10. 59 Table 13: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.5 F(X) BM NM PCM-1 PCM-2 NI x y z x Y z x Y z x y z 1 -0.0513 2.9983 -3.4472 -0.0513 2.9983 -3.4472 0.6441 0.9756 -3.1375 0.2857 2.1708 -3.2292 2 0.3572 1.7757 -3.3363 1.1912 -0.5845 -1.9070 0.8464 0.5118 0.5857 0.8117 -0.4824 0.8327 3 1.5987 -3.1866 -2.4532 1.1080 -0.2296 0.8874 1.0422 1.0307 0.9238 1.0437 1.0274 0.9220 4 -0.0691 -4.0377 -1.7322 1.1378 1.0118 0.8874 1.0421 1.0311 0.9238 1.0421 1.0311 0.9238 5 -0.5645 -3.1357 -1.4904 1.0421 1.0320 0.9239 1.0421 1.0311 0.9238 1.0421 1.0311 0.9238 6 -0.4704 -3.8333 -1.1061 1.0421 1.0311 0.9238 7 -0.2893 -3.6157 -0.8011 1.0421 1.0311 0.9238 8 2.9731 -0.1980 4.7720 9 2.9326 2.0404 3.9224 10 0.7844 1.0117 1.2238 11 0.9634 1.2176 0.8498 12 1.0141 1.0853 0.9036 13 1.0423 1.0308 0.9240 14 1.0415 1.0329 0.9219 15 1.0421 1.0311 0.9238 16 1.0421 1.0311 0.9238 As it is demonstrated in this table, the methods took different number of iterations until the process converged to the approximate solution. In general, the two predictor-corrector methods PCM-1 and PCM-2 are fast convergent methods when compared with that of Newton and Broyden’s methods as we have seen in all given examples. Both methods took a less number of iterations in all discussed examples. Example 7.6(Gilberto, 2004): Find the approximate solution for the following system of nonlinear equations if the initial approximation of the system is given by X (0) = (2 1)T . 𝑥12 + 𝑥22 − 50 = 0 𝑥1 . 𝑥2 − 25 = 0 60 Solution: This system of nonlinear equations has been solved using the same codes in appendices E, F, G and H. The only changes made for this example is the use of different SNLEs with different initial approximation value. Table 14: Iteration results of BM, NM, PCM-1 and PCM-2 for Example 7.6 F(x) BM NM PCM-1 PCM-2 NI 𝑥1 𝑥2 𝑥1 𝑥2 𝑥1 𝑥2 𝑥1 𝑥2 1 9.3333 8.8333 9.3333 8.8333 3.8163 3.4829 5.0640 4.6354 2 3.8772 3.4127 6.0428 5.7928 5.0176 4.9065 5.0714 4.9285 3 4.7441 4.3844 5.1337 5.0087 5.0185 4.9815 5.0238 4.9762 4 5.2024 49415 5.0317 4.9692 5.0062 4.9938 5.0079 4.9921 5 5.1149 4.8781 5.0156 4.9844 5.0021 4.9979 5.0026 4.9974 6 5.1042 4.8956 5.0078 4.9922 5.0007 4.9993 5.0009 4.9991 7 5.0585 4.9416 5.0039 4.9961 5.0002 4.9998 5.0003 4.9997 8 5.0375 4.9625 5.0020 4.9980 5.0000 5.0000 5.0000 5.0000 9 5.0228 4.9772 5.0010 4.9990 5.0000 5.0000 5.0000 5.0000 10 5.0142 4.9858 5.0005 4.9995 11 5.0087 4.9913 5.0002 4.9998 12 5.0054 4.9946 5.0000 5.0000 13 5.0033 4.9967 5.0000 5.0000 14 5.0021 4.9979 15 5.0013 4.9987 16 5.0008 4.9992 17 18 19 5.0005 4.9995 5.0003 4.9997 5.0002 4.9998 20 5.0000 5.0000 21 5.0000 5.0000 As it is demonstrated in the above table, the methods took different number of iterations until the process of iteration converged to the approximate solution [5.0000 5.0000]T . From this table Broyden took 21 iterations, Newton method took 13 iterations and both PCM-1 and 2 took 9 iterations. In general, the two predictor-corrector methods PCM-1 and PCM-2 are fast convergent methods when compared with that of Newton and Broyden’s methods as we have seen in all given examples. Both methods took a less number of iterations in all discussed examples. 61 8. SUMMARY, CONCLUSION AND RECOMMENDATION 8.1. Summary System of nonlinear equations arises in different areas of applied sciences such as engineering, physics, medicines, chemistry, and robotics. They appear also in many geometric computations such as intersections, minimum distance, and when solving initial or boundary value problems in ordinary or partial differential equations. Solving system of nonlinear equations is much more difficult when compared with solving the linear counterpart and finding solutions of these problems is one of the most important parts of the studies in numerical analysis. Because of this reason a number of mathematicians developed different iterative methods to solve problems of systems of nonlinear equations. Some important methods for solving system of nonlinear equations are methods like Newton’s method, Steepest Decent method, Broyden’s method, Predictor-corrector methods, etc are a few of them. Mainly, this project paper focused on Newton’s method, Broyden’s method and two iterative methods with two steps each (predictor-corrector methods 1 and 2) for solving system of nonlinear equations. The Predictor-corrector methods 1 and 2 are methods derived from closed-open quadrature formulas by using Taylor polynomial. The general aim of this project is to study iterative methods: Newton’s method, Broyden’s method, predictor-corrector method-1 and predictor-corrector method-2 which use to solve system of nonlinear equations and to compare the performance of all these methods one with the other based on the number of iterations which took until they reach to the solution. The iterative schemes for these methods has been given and discussed how to derive them. Some numerical examples have been given to indicate the performance of the methods and showed the number of iterations until the process come to converge to the approximate solution in a 4-decimal place rounding. From the comparison made among the given methods, the PCM-1 and PCM-2 were necessarily better performing methods than the others. 62 8.2. Conclusion In this project four iterative methods for solving system of nonlinear equations have been discussed in detail and the iterative schemes for each method are derived. These methods are Newton’s method, Broyden’s method, PCM-1 and PCM-2. The comparison among the four methods is made based on the number of iterations they took until they converge to the true solution and it is shown from all the test problems that PCM-1 and PCM-2 are better performing methods than that of the Newton and Broyden’s methods. This means that in all computations performed, it is PCM-1 and PCM-2 that took fewer numbers of iterations than Newton and Broyden’s methods took until converging to the solution. So this is the implication for both methods PCM-1 and PCM-2 to be more important to use them than other methods which are compared to them. Not only this, but also both PCM-1 and PCM-2 being cubic convergent makes them better than those Newton and Broyden’s methods for Newton’s method is quadratically convergent and Broyden’s method super linearly convergent. 8.3. Recommendation Based on the performance of the given methods, the following recommendations are suggested. From iterative methods for solving SNLEs some are NM, BM, PCM-1 and PCM-2. Among these methods, using PCM-1 and 2 help the user to reach to the solution after a few numbers of iterations when compared with NM and BM. Individuals interested to work on SNLEs using one of the four methods can benefit from the result and the MATLAB codes are presented in appendices to use by changing the necessary parts of the codes depending up on the size of the system of nonlinear equations they solve. It is necessary to show the way how to choose initial approximate values if individuals who need to study more on SNLEs to begin the iteration when they need to find a solution and it is also recommended to have a theory how a comparison is made between two or more iterative methods. 63 9. REFERENCES Andreas Griewank, 1986. The global convergence of Broyden-like methods with a suitable line search. Journal of Australia, Mathematics Soc. Ser. B, 28:75. Anuradha Singh and Jaiswal, J. P. 2013. Third and fourth-order iterative methods for solving systems of nonlinear Equations. Mathematics Subject Classification 2000: 65H10, 65Y20. Aslam Noor, M. 2007. Fifth-order convergent iterative method for solving nonlinear equations using quadrature formula. Journal of Mathematics Control Science Application, 1: 241-249. Aslam Noor, M. 2007. Some applications of quadrature formulas for solving nonlinear equations. Nonlinear Analysis Forum, 12 (1): 91-96. Atkinson, K. and Han, W. 2001. Theoretical Numerical Analysis. A Functional Analysis Framework. Springer-Verlag. Babajee, D.K.R., Dauhoo, M.Z., Darvishi, M.T. and Barati, A.2008. A note on the local convergence of iterative methods based on Adomian decomposition method and 3node quadrature rule. Applied Mathematics and Computation, 200: 452–458. Broyden,C.1965. A class of iterative methods for solving system of non-linear equations. Computational Mathematics, 19:577-593 Burden, R.L. and Faires, J.D. 2001. Numerical Analysis, 7th Edition. PWS Publishing Company, Boston. Burden, R.L., Faires, J.D. 2005. Numerical Solutions of Nonlinear Systems of Equations. Numerical Analysis, Belmount: Thomson Brooks/Cole (pp. 597-640). Chun, C. 2006. A new iterative method for solving nonlinear equations. Journal of applied mathematical and computation, 178(2): 415-422. Cordero, A. and Torregrosa, J.R. 2006. Variants of Newton’s method for functions of several variables. Applied Mathematics Computation, 183:199-208. Cordero, A. and Torregrosa, J.R. 2007. Variants of Newton’s method using fifth-order quadrature formulas. Applied Mathematics Computation, 190: 686_698. Courtney Remani, 2013. Numerical Methods for Solving Systems of Nonlinear Equations. Lakehead University, Thunder Bay, Ontario, Canada. 64 Daftardar-Gejji, V. and Jafari, H. 2006. An iterative method for solving nonlinear functional equations. Journal of Applied Mathematics, 316: 753–763. Darvishi, M. T. and Barati, A. 2007b. Super cubic iterative methods to solve systems nonlinear equations. Journal of applied mathematical and computation, 188: 16781685. Darvishi, M. T., and Barati A. 2007a. A third-order Newton-type method to solve systems of nonlinear equations. Journal of applied mathematical and computation, 187: 630-635. Darvishi, M.T. 2009. A two-step high order Newton-like method for solving systems of nonlinear equations. International Journal of Pure and Applied Mathematics, 57(4):543-555. Darvishi, M.T. and Barati, A. 2004. A fourth-order method from quadrature formulae to solve systems of nonlinear equations. Applied Mathematics Computation, 188: 257-261. Deuflhard, 2012. A short history of Newton's method. Documenta Mathematica, Optimization stories: 25-30. Edwards, Jr. 1997. The Historical Development of the Calculus. Springer-Verlag. Fosdick, L., Jessup, E., Schauble, C., and G. Domik, G., 1996. An Introduction to HighPerformance. Scientific Computing, MIT Press. Frontini, M. and Sormani, E. 2004. Third-order methods from quadrature formulae for solving systems of nonlinear equations. Applied Mathematics and Computation, 149: 771-782. Gautschi, W. 1997. Numerical Analysis: An Introduction, Birkhäuser. Gilberto Urroz E., 2004. Solution of non-linear equations. Applied Mathematics, 9-16. Golbabai, A., and Javidi, M. 2007. A new family of iterative methods for solving system of nonlinear algebraic equations. Journal of applied mathematical and computation, 190: 1717-1722. Griewank, A., 2012. Broyden updating, the good and the bad! Documenta Mathematica, Optimization stories, 301–315. Hafiz, M. A. and Mohamed Bahgat, S. M. 2012. An Efficient Two-step Iterative Method for Solving System of Nonlinear Equations. Journal of Mathematics Research, doi.org/10.5539/jmr.v4n4p28. Hafiz, M.A., and Bahgat, M.S.M., 2012. Extended and modified Halley's iterative method for solving nonlinear system of equations. Journal of Mathematical and Computational Sciences, 1(2):1-14. 65 Hassan, M. and Yusuf Waziri, M., 2015. On Broyden-like update via some quadrature formulas for solving nonlinear systems of equations. Turkish Journal of Mathematics, 39:335-345. Hosseini, S.M. 2010. Improving homotopy analysis method for system of nonlinear algebraic equations. Journal of Advanced Research in Applied Mathematics, 4 (2):22-30. Mamat, M., Kabiru Muhammad and Muhammad Waziri, Y. 2014. Trapezoidal Broyden’s method for solving systems of nonlinear equations, Applied Mathematical Sciences, 6 (8): 251 – 260. Martinez, J.M., 2000. Practical quasi-Newton methods for solving nonlinear systems. Journal of Computational and Applied Mathematics, 124:97-121. Michael Heath, T., 2002. Scientific computing: An introductory survey. Applied Mathematics Computation. University of Illinois at Urbana-Champaign, 48-53. Muhammad Aslam Noor, Muhammad Waseem and Khalida Inayat Noor, 2015. New iterative technique for solving a system of nonlinear equations. Applied Mathematics and Computation, 271: 446- 466. Noor, M.A. and Noor, K.I., 2006. Three-step iterative methods for nonlinear equations, Applied Mathematics and Computation, 183: 322–327. Noor, M.A. and Waseem, M., 2008. Iterative methods for solving a system of nonlinear equations. Computers and Mathematics with Applications, 57:101-106. Noor, M.A., Noor, K.I., Mohyud-Din, S.T. and Shabbir, A. 2006. An iterative method with cubic convergence for nonlinear equations, Applied Mathematics and Computation, 183:1249–1255. Ozel, M. 2010. A New Decomposition Method for Solving System of Nonlinear Equations. Journal of applied mathematical and computation, 15: 89-95. Podisuk, M., Chundang, U., Sanprasert, W. 2007. Single-step formulas and multi-step formulas of the integration method for solving the initial value problem of ordinary differential equation. Applied Mathematics Computation, 190: 1438-1444. Timothy Sauer, 2006. Numerical Analysis. Addison-Wesley, Boston, MA. Traub, J.F. 1964. Iterative Methods for the Solution of Equations. Prentice-Hall Englewood Cliffss, New Jersey, USA. Weerakoon, S. and Fernando, T.G.I. 2000. A variant of Newton.s method with accelerated third-order convergence. Applied Mathematics Lett. 13: 87-93. 66 10. APPENDICES The following are MATLAB codes for Newton, PCM-1, PCM-2 and Broyden’s methods. Appendix- A MATLAB code for Newton’s method: File name: nm3x3.m %Newton method to solve nonlinear system of equations f(X)=0 with three equations and three unknowns %input f and x0 %x0-initial guess a column vector in R^3 %e.g. x0=[0.1 0.1 -0.1]'; %f-a column vector valued function from R^3 into R^3 %e.g. f=[3*x-cos(y*z)-0.5;x^2-81*(y+0.1)^2+sin(z)+1.06;exp(-x*y)+20*z+((10*sym(pi)3)/3)]; function nm3x3(f,x0) syms x y z maxiter=100; jb=jacobian(f,[x y z]); A=subs(f,[x y z],x0); B=subs(jb,[x y z],x0);%the jacobian matrix of f evaluated at initial guess x0 x1=x0-inv(B)*A; z1=ones(3,100); z1(1:3,1)=x0; z1(1:3,2)=x1; k=1; while ((norm(z1(1:3,k+1)-z1(1:3,k),inf)>=10^-4)&(k<=maxiter)) A=subs(f,[x y z],z1(1:3,k+1)); B=subs(jb,[x y z],z1(1:3,k+1)); z1(1:3,k+2)=z1(1:3,k+1)-inv(B)*A; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:3,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:3,j+1)-z1(1:3,j))) end disp('Solution='),disp(z1(1:3,k+1)) 67 soln=subs(f,[x y z],z1(1:3,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end Appendix- B MATLAB code for PCM-1 File name: pcm-1.m %predictor corrector method-1 to solve nonlinear system of equations f(X) =0 with three equations and three unknowns %input f and x0 %x0-initial guess a column vector in R^3 %e.g. x0=[0.1 0.1 -0.1]'; %f-a column vector valued function from R^3 into R^3 %e.g. f=[3*x-cos(y*z)-0.5;x^2-81*(y+0.1)^2+sin(z)+1.06;exp(-x*y)+20*z+((10*sym(pi)3)/3)]; function pcm_1(f,x0) syms x y z maxiter=100; jb=jacobian(f,[x y z]); fx0=subs(f,[x y z],x0); jbx0=subs(jb,[x y z],x0); y0=x0-inv(jbx0)*fx0; jbxy0=subs(jb,[x y z],(x0+2*y0)/3); x1=x0-4*inv(jbx0+3*jbxy0)*fx0; w1=ones(3,100); w1(1:3,1)=y0; z1=ones(3,100); z1(1:3,1)=x0; z1(1:3,2)=x1; k=1; while ((norm(z1(1:3,k+1)-z1(1:3,k))>=10^-4)&(k<=maxiter)) A=subs(f,[x y z],z1(1:3,k+1)); B=subs(jb,[x y z],z1(1:3,k+1)); w1(1:3,k+1)=z1(1:3,k+1)-inv(B)*A; C=subs(jb,[x y z],(z1(1:3,k+1)+2*w1(1:3,k+1))/3); z1(1:3,k+2)=z1(1:3,k+1)-4*inv(B+3*C)*A; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) 68 iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:3,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:3,j+1)-z1(1:3,j))) end disp('Solution='),disp(z1(1:3,k+1)) soln=subs(f,[x y z],z1(1:3,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end Appendix-C MATLAB code for PCM-2 File name: pcm-2.m %predictor corrector method-2 to solve nonlinear system of equations f(X)=0 with three equations and three unknowns %input f and x0 %x0-initial guess a column vector in R^3 %e.g. x0=[0.1 0.1 -0.1]'; %f-a column vector valued function from R^3 into R^3 %e.g. f=[3*x-cos(y*z)-0.5;x^2-81*(y+0.1)^2+sin(z)+1.06;exp(-x*y)+20*z+((10*sym(pi)3)/3)]; function pcm_2(f,x0) syms x y z maxiter=100; jb=jacobian(f,[x y z]); fx0=subs(f,[x y z],x0); jbx0=subs(jb,[x y z],x0); y0=x0-inv(jbx0)*fx0; jbxy0=subs(jb,[x y z],(2*x0+y0)/3); x1=x0-4*inv(jbx0+3*jbxy0)*fx0; w1=ones(3,100); w1(1:3,1)=y0; z1=ones(3,100); z1(1:3,1)=x0; z1(1:3,2)=x1; k=1; while ((norm(z1(1:3,k+1)-z1(1:3,k))>=10^-4)&(k<=maxiter)) A=subs(f,[x y z],z1(1:3,k+1)); B=subs(jb,[x y z],z1(1:3,k+1)); w1(1:3,k+1)=z1(1:3,k+1)-inv(B)*A; B1=subs(jb,[x y z],w1(1:3,k+1)); 69 C=subs(jb,[x y z],(2*z1(1:3,k+1)+w1(1:3,k+1))/3); z1(1:3,k+2)=z1(1:3,k+1)-4*inv(3*C+B1)*A; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:3,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:3,j+1)-z1(1:3,j))) end disp('Solution='),disp(z1(1:3,k+1)) soln=subs(f,[x y z],z1(1:3,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end Appendix-D MATLAB code for Broyden’s method File name: broyden3x3.m The code is given by: %Broyden method to solve nonlinear system of equations f(X)=0 with three equations and three unknowns %input f and x0 %x0-initial guess a column vector in R^3 %e.g. x0=[0.1 0.1 -0.1]'; %f-a column vector valued function from R^3 into R^3 %e.g. f=[3*x-cos(y*z)-0.5;x^2-81*(y+0.1)^2+sin(z)+1.06;exp(-x*y)+20*z+((10*sym(pi)3)/3)]; function broyden3x3(f,x0) syms x y z maxiter=100; jb=jacobian(f,[x y z]); B0=subs(jb,[x y z],x0); d0=-inv(B0)*subs(f,[x y z],x0); x1=x0+d0; z1=ones(3,100); z1(1:3,1)=x0; z1(1:3,2)=x1; 70 Bk=B0; dk=d0; k=1; while ((norm(z1(1:3,k+1)-z1(1:3,k))>=10^-4)&(k<=maxiter)) A=subs(f,[x y z],z1(1:3,k+1))*dk'; c=dk'*dk; Bk=Bk+(A/c); dk=-inv(Bk)*subs(f,[x y z],z1(1:3,k+1)); z1(1:3,k+2)=z1(1:3,k+1)+dk; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:3,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:3,j+1)-z1(1:3,j))) end disp('Solution='),disp(z1(1:3,k+1)) soln=subs(f,[x y z],z1(1:3,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end Appendix- E MATLAB Codes Broyden2x2.m %Broyden method to solve nonlinear system of equations f(X)=0 with two equations and two unknowns %input f and x0 %x0-initial guess a column vector in R^2 %e.g. x0=[0.1 0.1]'; %f-a column vector valued function from R^2 into R^2 %e.g. f=[x^2-81*(y+0.1)^2+sin(z)+1.06;exp(-x*y)+20*z+((10*sym(pi)-3)/3)]; function broyden2x2(f,x0) syms x y maxiter=100; jb=jacobian(f,[x y]); B0=subs(jb,[x y],x0); d0=-inv(B0)*subs(f,[x y],x0); 71 x1=x0+d0; z1=ones(2,100); z1(1:2,1)=x0; z1(1:2,2)=x1; Bk=B0; dk=d0; k=1; while ((norm(z1(1:2,k+1)-z1(1:2,k))>=10^-4)&(k<=maxiter)) A=subs(f,[x y],z1(1:2,k+1))*dk'; c=dk'*dk; Bk=Bk+(A/c); dk=-inv(Bk)*subs(f,[x y],z1(1:2,k+1)); z1(1:2,k+2)=z1(1:2,k+1)+dk; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:2,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:2,j+1)-z1(1:2,j))) end disp('Solution='),disp(z1(1:2,k+1)) soln=subs(f,[x y],z1(1:2,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end Appendix- F MATLAB Codes nm2x2.m %Newton method to solve nonlinear system of equations f(X)=0 with two equations and two unknowns %input f and x0 %x0-initial guess a column vector in R^2 %e.g. x0=[0.1 0.1]'; %f-a column vector valued function from R^2 into R^2 %e.g. f=[3*x-cos(y)-0.5;x^2-81*(y+0.1)^2+1.06]; function nm2x2(f,x0) syms x y 72 maxiter=100; jb=jacobian(f,[x y]); A=subs(f,[x y],x0); B=subs(jb,[x y],x0);%the jacobian matrix of f evaluated at initial guess x0 x1=x0-inv(B)*A; z1=ones(2,100); z1(1:2,1)=x0; z1(1:2,2)=x1; k=1; while ((norm(z1(1:2,k+1)-z1(1:2,k))>=10^-4)&(k<=maxiter)) A=subs(f,[x y],z1(1:2,k+1)); B=subs(jb,[x y],z1(1:2,k+1)); z1(1:2,k+2)=z1(1:2,k+1)-inv(B)*A; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:2,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:2,j+1)-z1(1:2,j))) end disp('Solution='),disp(z1(1:2,k+1)) soln=subs(f,[x y],z1(1:2,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end Appendix- G MATLAB Codes Pcml.m %predictor corrector method-1 to solve nonlinear system of equations f(X)=0 with two equations and two unknowns %input f and x0 %x0-initial guess a column vector in R^2 %e.g. x0=[0.1 0.1]'; %f-a column vector valued function from R^2 into R^2 %e.g. f=[x^2-81*(y+0.1)^2+sin(z)+1.06;exp(-x*y)+20*z+((10*sym(pi)-3)/3)]; function pcm1(f,x0) syms x y 73 maxiter=100; jb=jacobian(f,[x y]); fx0=subs(f,[x y],x0); jbx0=subs(jb,[x y],x0); y0=x0-inv(jbx0)*fx0; jbxy0=subs(jb,[x y],(x0+2*y0)/3); x1=x0-4*inv(jbx0+3*jbxy0)*fx0; w1=ones(2,100); w1(1:2,1)=y0; z1=ones(2,100); z1(1:2,1)=x0; z1(1:2,2)=x1; k=1; while ((norm(z1(1:2,k+1)-z1(1:2,k))>=10^-4)&(k<=maxiter)) A=subs(f,[x y],z1(1:2,k+1)); B=subs(jb,[x y],z1(1:2,k+1)); w1(1:2,k+1)=z1(1:2,k+1)-inv(B)*A; C=subs(jb,[x y],(z1(1:2,k+1)+2*w1(1:2,k+1))/3); z1(1:2,k+2)=z1(1:2,k+1)-4*inv(B+3*C)*A; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:2,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:2,j+1)-z1(1:2,j))) end disp('Solution='),disp(z1(1:2,k+1)) soln=subs(f,[x y],z1(1:2,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end Appendix- H MATLAB Codes Pcm2.m %predictor corrector method-2 to solve nonlinear system of equations f(X)=0 with two equations and two unknowns %input f and x0 74 %x0-initial guess a column vector in R^2 %e.g. x0=[0.1 0.1]'; %f-a column vector valued function from R^2 into R^2 %e.g. f=[x^2-81*(y+0.1)^2+sin(z)+1.06;exp(x*y)+20*z+((10*sym(pi)-3)/3)]; function pcm2(f,x0) syms x y maxiter=100; jb=jacobian(f,[x y]); fx0=subs(f,[x y],x0); jbx0=subs(jb,[x y],x0); y0=x0-inv(jbx0)*fx0; jbxy0=subs(jb,[x y],(2*x0+y0)/3); x1=x0-4*inv(jbx0+3*jbxy0)*fx0; w1=ones(2,100); w1(1:2,1)=y0; z1=ones(2,100); z1(1:2,1)=x0; z1(1:2,2)=x1; k=1; while ((norm(z1(1:2,k+1)-z1(1:2,k))>=10^-4)&(k<=maxiter)) A=subs(f,[x y],z1(1:2,k+1)); B=subs(jb,[x y],z1(1:2,k+1)); w1(1:2,k+1)=z1(1:2,k+1)-inv(B)*A; B1=subs(jb,[x y],w1(1:2,k+1)); C=subs(jb,[x y],(2*z1(1:2,k+1)+w1(1:2,k+1))/3); z1(1:2,k+2)=z1(1:2,k+1)-4*inv(3*C+B1)*A; k=k+1; end if k>maxiter disp('too many iterations') else disp('Number of iteration k=') disp(k) iter=k; disp('value of X(j) at iteration j where j=0,1,2,...,k') disp(z1(1:2,1:k+1)) disp('stopping condition ||X(j)-X(j-1)|| for j=1,...,k') for j=1:k disp(norm(z1(1:2,j+1)-z1(1:2,j))) end disp('Solution='),disp(z1(1:2,k+1)) soln=subs(f,[x y],z1(1:2,k+1)); disp('check for solution') disp('f[X(k)]=') disp(soln) end