Efficient calculation of the Gauss-Newton approximation of the Hessian matrix in neural networks.

Michael Fairbank, Eduardo Alonso
Author Information
  1. Michael Fairbank: Department of Computing, School of Informatics, City University London, London EC1V 0HB, U.K. michael.fairbank1@city.ac.uk

Abstract

The Levenberg-Marquardt (LM) learning algorithm is a popular algorithm for training neural networks; however, for large neural networks, it becomes prohibitively expensive in terms of running time and memory requirements. The most time-critical step of the algorithm is the calculation of the Gauss-Newton matrix, which is formed by multiplying two large Jacobian matrices together. We propose a method that uses backpropagation to reduce the time of this matrix-matrix multiplication. This reduces the overall asymptotic running time of the LM algorithm by a factor of the order of the number of output nodes in the neural network.

MeSH Term

Algorithms
Neural Networks, Computer

Word Cloud

Created with Highcharts 10.0.0algorithmneuralnetworkstimeLMlargerunningcalculationGauss-NewtonmatrixLevenberg-Marquardtlearningpopulartraininghoweverbecomesprohibitivelyexpensivetermsmemoryrequirementstime-criticalstepformedmultiplyingtwoJacobianmatricestogetherproposemethodusesbackpropagationreducematrix-matrixmultiplicationreducesoverallasymptoticfactorordernumberoutputnodesnetworkEfficientapproximationHessian

Similar Articles

Cited By (1)