Multi-spectral gradient method via variational technique under log-determinant norm for large-scale optimization

The spectral gradient method is popular due to the fact that only the gradient of the objective function is required at each iterate. Besides that, it is more efficient than the quasi-Newton method as the storage of second derivatives (Hessian) approximation are not required especially when the dime...

詳細記述

保存先:
書誌詳細
主要な著者: Hong, Seng Sim, Leong, Wah June, Chen, Chuei Yee, Ibrahim, Siti Nur Iqmal
フォーマット: 論文
言語:English
出版事項: Malaysian Mathematical Science Society 2017
オンライン・アクセス:http://psasir.upm.edu.my/id/eprint/62502/1/SPECTRAL.pdf
タグ: タグ追加
タグなし, このレコードへの初めてのタグを付けませんか!
その他の書誌記述
要約:The spectral gradient method is popular due to the fact that only the gradient of the objective function is required at each iterate. Besides that, it is more efficient than the quasi-Newton method as the storage of second derivatives (Hessian) approximation are not required especially when the dimension of the problem is large. In this paper, we propose a spectral gradient method via variational technique under log-determinant measure such that it satisfies the weaker secant equation. The corresponding variational problem is solved and the Lagrange multiplier is approximated using the Newton-Raphson method and solved following interior point method that is associated with weaker secant relation. An executable code is developed to test the efficiency of the proposed method with some standard conjugate-gradient methods. Numerical results are presented which suggest a better performance has been achieved.