A family of gradient based on supervised training algorithm and its application is discussed. A Levenberg-Marquardt (LM) algorithm, a combination of the gradient method and the Gauss-Newton method, is under close investigation. It boasts of the local convergence of the Gauss-Newton method as well as the overall convergence of the gradient method. When μ is large, the LM algorithm looks like the gradient method; when μ is small, it changes into the Gauss-Newton method. With the aid of the approximate second derivative, the LM algorithm is more efficient than the gradient method. As to with the training process and accuracy, the LM method is superior to the conjugate gradient method and the vary-learning-rate BP method. As the main matrix is positively definite, the solution always exists. From this view point, the LM method is preferable to the Gauss-Newton method.
|Original language||English (US)|
|Number of pages||3|
|Journal||Qinghua Daxue Xuebao/Journal of Tsinghua University|
|State||Published - 1997|
ASJC Scopus subject areas
- Computer Science Applications
- Applied Mathematics