Description: 这里实现了一个最优化控制的算法,牛顿梯度法的源代码,运行于matlab平台下。-here to achieve optimization of a control algorithm, Newton gradient of the source code, which runs on Matlab platforms. Platform: |
Size: 1024 |
Author:周行星 |
Hits:
Description: 多变量寻优的源码,包括基本算法如下:DFP法(又称变尺度法),BFS法、一阶梯度法,共扼梯度法。-multivariate optimization of the source code, including basic algorithm is as follows : DFP Act (also known as the variable scale), BFS, a ladder degrees, a total of the accused gradient method. Platform: |
Size: 4096 |
Author:陈镇静 |
Hits:
Description: optimisation中的梯度法,在实验课上编写的程序,可以直接看到效果-optimization of the gradient method, in the experimental procedure for the preparation of class, you can directly see the effect of Platform: |
Size: 1024 |
Author:bellepdt |
Hits:
Description: 一种带有梯度加速的粒子群算法,可以实现多种优化工作的需要哦。-With a gradient to accelerate the particle swarm algorithm, can achieve the needs of a variety of optimization Oh. Platform: |
Size: 34816 |
Author:qubin |
Hits:
Description: 约束最优化方法--最速下降法(也叫梯度法),是人们用来求多个变量函数极值问题的最早的一种方法。-Constrained optimization methods- steepest descent method (also known as gradient method), is used for multiple variables function Extremum Problems earliest methods. Platform: |
Size: 2048 |
Author:anytry |
Hits:
Description: 梯度寻优,二阶模型的辨识,结果很准,工业控制热工传递曲线等-Gradient optimization, second-order model identification, the result is quasi- Platform: |
Size: 1024 |
Author:崔到平 |
Hits:
Description: 这是一个图像刚性配准的程序,有matlab的界面,能够应用pv内插算法,NMI测度和共轭梯度优化算法实现-this is a image registration code of matlab。 it include pv interpretation,NMI metric and conjugate gradient optimization。
Platform: |
Size: 105472 |
Author:xuhui |
Hits:
Description: 压缩包里包含了无约束优化问题常用的几种求解方法的源程序:变量轮换法(variable_rotation.m)、最速下降法(steepest_descent.m)、修正牛顿法(modified_newton.m)、共轭梯度法(conjugate_gradient.m)。另外,coefficient_matrix.m为目标函数系数获得矩阵,minval.m为最小值计算函数,gradient.m为梯度计算函数-Compression bag contains unconstrained optimization problems of several commonly used method of source: variable rotation Act (variable_rotation.m), steepest descent method (steepest_descent.m), as amended Newton (modified_newton.m), conjugate gradient method (conjugate_gradient.m). In addition, coefficient_matrix.m as the objective function coefficients obtained matrix, minval.m function for the minimum calculation, gradient.m for gradient calculation function-Compression bag contains unconstrained optimization problems of several commonly used method of source: variable rotation Act (variable_rotation.m), steepest descent method (steepest_descent.m), as amended Newton (modified_newton.m), conjugate gradient method (conjugate_gradient.m). In addition, coefficient_matrix.m as the objective function coefficients obtained matrix, minval.m function for the minimum calculation, gradient.m for gradient calculation function Platform: |
Size: 1024 |
Author:zhuyuanli |
Hits:
Description: 通过使用matlab求解共轭梯度法和牛顿法。熟悉经典优化方法。(Solve the conjugate gradient method and the Newton method by using matlab. Familiar with the classic optimization method.) Platform: |
Size: 1024 |
Author:亮看世界
|
Hits:
Description: 基于Armijo-Goldstein准则的用matlab实现的共轭梯度优化方法,个人编写,适合优化方法入门练习。(Based on the Armijo-Goldstein criterion, the conjugate gradient optimization method implemented by Matlab is written by individual, suitable for the introduction of the optimization method) Platform: |
Size: 1024 |
Author:Victor_Z |
Hits:
Description: 遗传算法提供了求解非线性规划的通用框架,它不依赖于问题的具体领域。遗传算法的优点是将问题参数编码成染色体后进行优化, 而不针对参数本身, 从而不受函数约束条件的限制; 搜索过程从问题解的一个集合开始, 而不是单个个体, 具有隐含并行搜索特性, 可大大减少陷入局部最小的可能性。而且优化计算时算法不依赖于梯度信息,且不要求目标函数连续及可导,使其适于求解传统搜索方法难以解决的大规模、非线性组合优化问题。(Genetic algorithm provides a general framework for solving nonlinear programming, which does not depend on the specific problem domain. The advantage of genetic algorithm is that the problem parameters are encoded into chromosomes for optimization, rather than the parameters themselves. The search process starts from a set of problem solutions, rather than a single individual, and has the implicit parallel search feature, which can greatly reduce the possibility of falling into the local minimum. Moreover, the algorithm does not rely on gradient information and does not require the objective function to be continuous and differentiable, which makes it suitable for solving large-scale and nonlinear combinatorial optimization problems that are difficult to be solved by traditional search methods.) Platform: |
Size: 33792 |
Author:FZenjoys |
Hits: