Welcome![Sign In][Sign Up]
Location:
Search - conjugate inverse

Search list

[matlabgetd

Description: 共轭梯度法,是数值分析中很重要的一种,源码为其在matlab中的实现。-Conjugate gradient method, numerical analysis is a very important one in the matlab source code for its realization.
Platform: | Size: 1024 | Author: 马丫 | Hits:

[matlabcgls

Description: 用于解反问题的共轭梯度法,对于Ax=b,输入矩阵A,列向量b,以及迭代步数k,可求的列向量x-Solution of inverse problems for the conjugate gradient method, for Ax = b, the input matrix A, the column vector b, as well as the number of iterations k, rectifiable column vector x
Platform: | Size: 1024 | Author: gongwei | Hits:

[matlabquaternion

Description: 四元数乘法、求逆、共轭、求范数函数,并附有求解矢量旋转坐标的程序例子-Quaternion multiplication, inverse, conjugate, seeking norm function, together with procedures for solving the example of vector rotating coordinate
Platform: | Size: 2048 | Author: wbw | Hits:

[matlabConjugate-gradient

Description: 共轭梯度法是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。-Conjugate gradient method is between the steepest descent method and Newton method between a method that only use the first derivative information, but the steepest descent method to overcome the disadvantage of slow convergence, but also avoids the need to store and calculate Newton Hesse matrix and the shortcomings of the inverse, conjugate gradient method is not only linear equations to solve large-scale one of the most useful, large-scale nonlinear optimization solution is also the most efficient algorithms.
Platform: | Size: 1024 | Author: sunling | Hits:

[AlgorithmNewton

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。 在各种优化算法中,共轭梯度法是非常重要的一种。其优点是所需存储量小,具有步收敛性,稳定性高,而且不需要任何外来参数。-Conjugate Gradient method (Conjugate Gradient) is between the steepest descent method between Newton method and a method, it only USES a derivative information, but overcome the steepest descent method slow convergence of weakness, but also avoid the Newton law needs to storage and computing Hesse inverse matrix and shortcomings, Conjugate Gradient method is not only solve linear equations with most of the large method, and also one of the most effective solution large nonlinear optimization of one of the algorithm. In all kinds of optimization algorithm, the conjugate gradient method is very important. Its advantage is the storage capacity needed, it has small step convergence, high stability, and doesn t require any exotic parameters.
Platform: | Size: 1024 | Author: 兰中周 | Hits:

[matlabgongetidufadshuzhixingzhi

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。 在各种优化算法中,共轭梯度法是非常重要的一种。其优点是所需存储量小,具有步收敛性,稳定性高,而且不需要任何外来参数-Conjugate Gradient method (Conjugate Gradient) is between the steepest descent method between Newton method and a method, it only USES a derivative information, but overcome the steepest descent method slow convergence of weakness, but also avoid the Newton law needs to storage and computing Hesse inverse matrix and shortcomings, Conjugate Gradient method is not only solve linear equations with most of the large method, and also one of the most effective solution large nonlinear optimization of one of the algorithm. In all kinds of optimization algorithm, the conjugate gradient method is very important. Its advantage is the storage capacity needed, it has small step convergence, high stability, and doesn t require any exotic parameters numerical experiment, this is the modern scientific computing of the answer above problem sets
Platform: | Size: 1024 | Author: 兰中周 | Hits:

[source in ebookConjugate-Gradient-Method

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。-Conjugate gradient method (Conjugate Gradient) between the steepest descent between law and Newton' s method is a method, it is only the first derivative information, but to overcome the steepest descent method of slow convergence shortcomings, but also avoid the Newton method needs to be stored and calculate the Hesse matrix and the inverse of the shortcomings of the conjugate gradient method is not only the most useful way to solve the large linear equations, one is also the solution of large-scale nonlinear optimization one of the most effective algorithm.
Platform: | Size: 704512 | Author: | Hits:

[Algorithmirfft

Description: 计算共轭对称复序列的快速傅里叶反变换,其变换结果是实数。-Symmetric complex conjugate calculation sequence of the fast Fourier inverse transform, the transformed result is a real number.
Platform: | Size: 142336 | Author: 冰河 | Hits:

[matlabconjgradmethod

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。这里给出共轭梯度法的源程序-Conjugate gradient method (Conjugate Gradient) is between the steepest descent method and Newton' s method between a method that takes only a first derivative information, but to overcome the slow convergence of the steepest descent method shortcomings, but also avoid the need to store Newton and computing the inverse Hesse matrix and disadvantages, conjugate gradient method is not only to solve large linear equations of the most useful methods, large-scale nonlinear optimization solution is the most efficient algorithm. Here is the source conjugate gradient method
Platform: | Size: 1024 | Author: lucy | Hits:

[Algorithmbycgste

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一-Conjugate gradient method (Conjugate Gradient) is between the steepest descent method and Newton' s method between a method that takes only a first derivative information, but to overcome the slow convergence of the steepest descent method shortcomings, but also avoid the need to store Newton and computing the inverse Hesse matrix and disadvantages, conjugate gradient method is not only to solve large linear equations of the most useful methods, large-scale nonlinear optimization solution is the most efficient algorithms
Platform: | Size: 2048 | Author: wangwenshu | Hits:

[CSharptidu

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。-Conjugate Gradient Method (Conjugate Gradient) is between the steepest descent method between a law and Newton' s method, it is only the first order derivative information, but the steepest descent method overcomes the shortcomings of slow convergence and avoid the need to store the Newton law Hesse and cons of computing the inverse matrix and the conjugate gradient method is not only one of the most useful methods to solve large linear equations, but also large-scale nonlinear optimization solution of one of the most effective algorithm.
Platform: | Size: 1024 | Author: cheng | Hits:

[Algorithmivtcg

Description: 不完全变量截断共轭梯度算法,目标函数的最优化求解,常用于逆问题的求解,是典型的L1范数算法。-Incomplete variable truncated conjugate gradient algorithm for solving optimization objective function, commonly used in inverse problem solving, is typical of L1-norm algorithm.
Platform: | Size: 2048 | Author: 张旭 | Hits:

[AlgorithmConjugate-Gradient-Method

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。 在各种优化算法中,共轭梯度法是非常重要的一种。其优点是所需存储量小,具有步收敛性,稳定性高,而且不需要任何外来参数。-Conjugate gradient method (Conjugate Gradient) is between the steepest descent method between the method and Newton' s method, it takes only a first derivative information, but to overcome the steepest descent method convergence slow shortcomings, but also to avoid the Newton method needs to be stored Hesse and disadvantages of computing inverse matrix and the conjugate gradient method is not only one of the most useful methods to solve large linear equations, solution of large-scale nonlinear optimization is one of the most effective algorithm. In various optimization algorithm, conjugate gradient method is a very important one. The advantage is that a small amount of memory required, with step convergence, high stability, and does not require any external parameters.
Platform: | Size: 367616 | Author: 陈怀兵 | Hits:

[matlabFR

Description: 共轭梯度法(Conjugate Gradient)是介于最速下降法与牛顿法之间的一个方法,它仅需利用一阶导数信息,但克服了最速下降法收敛慢的缺点,又避免了牛顿法需要存储和计算Hesse矩阵并求逆的缺点,共轭梯度法不仅是解决大型线性方程组最有用的方法之一,也是解大型非线性最优化最有效的算法之一。 在各种优化算法中,共轭梯度法是非常重要的一种。其优点是所需存储量小,具有步收敛性,稳定性高,而且不需要任何外来参数。-The Conjugate Gradient method is a method between the steepest descent method and the Newton method. It only needs to use the first derivative information, but overcomes the shortcomings of the steepest descent method and avoids the need for the Newton method to store And the calculation of Hesse matrix and the inverse of the shortcomings, conjugate gradient method is not only to solve large-scale linear equations one of the most useful methods, but also solution for large-scale nonlinear optimization of one of the most effective algorithm. In a variety of optimization algorithms, conjugate gradient method is a very important one. The advantage is that the required storage capacity is small, with step convergence, high stability, and does not require any external parameters.
Platform: | Size: 1024 | Author: 刘杉 | Hits:

[matlabEITtext

Description: 解决EIT中的逆问题,包括吉洪诺夫,Landewer,L1,共轭梯度法等(Solve the inverse problem in EIT, including gihunov, landewer, L1, conjugate gradient method and so on)
Platform: | Size: 43692032 | Author: 金刚father | Hits:

CodeBus www.codebus.net