z-logo
open-access-imgOpen Access
An Advanced Conjugate Gradient Training Algorithm Based on a Modified Secant Equation
Author(s) -
Ioannis E. Livieris,
Panagiotis Pintelas
Publication year - 2011
Publication title -
isrn artificial intelligence
Language(s) - English
Resource type - Journals
eISSN - 2090-7443
pISSN - 2090-7435
DOI - 10.5402/2012/486361
Subject(s) - conjugate gradient method , nonlinear conjugate gradient method , robustness (evolution) , gradient descent , artificial neural network , conjugate residual method , computer science , derivation of the conjugate gradient method , algorithm , gradient method , mathematics , line search , mathematical optimization , artificial intelligence , radius , biochemistry , chemistry , computer security , gene
Conjugate gradient methods constitute excellent neural network training methods characterized by their simplicity, numerical efficiency, and their very low memory requirements. In this paper, we propose a conjugate gradient neural network training algorithm which guarantees sufficient descent using any line search, avoiding thereby the usually inefficient restarts. Moreover, it achieves a high-order accuracy in approximating the second-order curvature information of the error surface by utilizing the modified secant condition proposed by Li et al. (2007). Under mild conditions, we establish that the proposed method is globally convergent for general functions under the strong Wolfe conditions. Experimental results provide evidence that our proposed method is preferable and in general superior to the classical conjugate gradient methods and has a potential to significantly enhance the computational efficiency and robustness of the training process.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom