Nonmonotone spectral gradient method based on memoryless symmetric rank-one update for large-scale unconstrained optimization
This paper proposes a nonmonotone spectral gradient method for solving large-scale unconstrained optimization problems. The spectral parameter is derived from the eigenvalues of an optimally sized memoryless symmetric rank-one matrix obtained under the measure defined as a ratio of the determinant...
保存先:
主要な著者: | , , , |
---|---|
フォーマット: | 論文 |
出版事項: |
American Institute of Mathematical Sciences
2021
|
オンライン・アクセス: | http://psasir.upm.edu.my/id/eprint/94373/ https://www.aimsciences.org/article/doi/10.3934/jimo.2021143 |
タグ: |
タグ追加
タグなし, このレコードへの初めてのタグを付けませんか!
|
要約: | This paper proposes a nonmonotone spectral gradient method for
solving large-scale unconstrained optimization problems. The spectral parameter is derived from the eigenvalues of an optimally sized memoryless symmetric
rank-one matrix obtained under the measure defined as a ratio of the determinant of updating matrix over its largest eigenvalue. Coupled with a nonmonotone line search strategy where backtracking-type line search is applied
selectively, the spectral parameter acts as a stepsize during iterations when no
line search is performed and as a milder form of quasi-Newton update when
backtracking line search is employed. Convergence properties of the proposed
method are established for uniformly convex functions. Extensive numerical
experiments are conducted and the results indicate that our proposed spectral
gradient method outperforms some standard conjugate-gradient methods. |
---|