
In mathematics, the conjugate gradient method is an algorithm for the numerical solution of particular systems of linear equations, namely those whose matrix is symmetric and positivedefinite. The conjugate gradient method is often implemented as an iterative algorithm, applicable to sparse systems that are too large to be handled by a direct implementation or other direct methods such as the Cholesky decomposition. Large sparse systems often arise when numerically solving partial differential equations or optimization problems. The conjugate gradient method can also be used to solve unconstrained optimization problems such as energy minimization. It was mainly developed by Magnus Hestenes and Eduard Stiefel. The biconjugate gradient method provides a generalization to nonsymmetric matrices. Various nonlinear conjugate gradient methods seek minima of nonlinear equations. ==Description of the method== Suppose we want to solve the following system of linear equations :Ax = b for the vector x where the known ''n'' × ''n'' matrix A is symmetric (i.e., A^{T} = A), positive definite (i.e. x^{T}Ax > 0 for all nonzero vectors x in R^{''n''}), and real, and b is known as well. We denote the unique solution of this system by . 抄文引用元・出典: フリー百科事典『 ウィキペディア（Wikipedia）』 ■ウィキペディアで「conjugate gradient method」の詳細全文を読む スポンサード リンク
