Neural Networks, IEEE - INNS - ENNS International Joint Conference on
Download PDF

Abstract

The paper describes a new training algorithm that has scalable memory requirements, which may range horn O(W) to O(W2), although in practice the useful range is limited to lower complexity levels. The algorithm is based upon a novel iterative estimation of the principal eigen-subspace of the Hessian, together with a quadratic step estimation procedure. It is shown that the new algorithm has convergence time comparable to conjugate gradient descent and maybe preferable if early stopping is used as it converges more quickly during the initial phases. Section 2 overviews the principles of second order training algorithms. Section 3 introduces the new algorithm. Second 4 discusses some experiments to confirm the algorithm's performance; section 5 concludes the paper.
Like what you’re reading?
Already a member?
Get this article FREE with a new membership!