Multi-spectral gradient method via variational technique under log-determinant norm for large-scale optimization

The spectral gradient method is popular due to the fact that only the gradient of the objective function is required at each iterate. Besides that, it is more efficient than the quasi-Newton method as the storage of second derivatives (Hessian) approximation are not required especially when the dime...

Celý popis

Uloženo v:
Podrobná bibliografie
Hlavní autoři: Hong, Seng Sim, Leong, Wah June, Chen, Chuei Yee, Ibrahim, Siti Nur Iqmal
Médium: Článek
Jazyk:English
Vydáno: Malaysian Mathematical Science Society 2017
On-line přístup:http://psasir.upm.edu.my/id/eprint/62502/1/SPECTRAL.pdf
Tagy: Přidat tag
Žádné tagy, Buďte první, kdo otaguje tento záznam!
Popis
Shrnutí:The spectral gradient method is popular due to the fact that only the gradient of the objective function is required at each iterate. Besides that, it is more efficient than the quasi-Newton method as the storage of second derivatives (Hessian) approximation are not required especially when the dimension of the problem is large. In this paper, we propose a spectral gradient method via variational technique under log-determinant measure such that it satisfies the weaker secant equation. The corresponding variational problem is solved and the Lagrange multiplier is approximated using the Newton-Raphson method and solved following interior point method that is associated with weaker secant relation. An executable code is developed to test the efficiency of the proposed method with some standard conjugate-gradient methods. Numerical results are presented which suggest a better performance has been achieved.