Effective neural network training with adaptive learning rate based on training loss.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

A method that uses an adaptive learning rate is presented for training neural networks. Unlike most conventional updating methods in which the learning rate gradually decreases during training, the proposed method increases or decreases the learning rate adaptively so that the training loss (the sum of cross-entropy losses for all training samples) decreases as much as possible. It thus provides a wider search range for solutions and thus a lower test error rate. The experiments with some well-known datasets to train a multilayer perceptron show that the proposed method is effective for obtaining a better test accuracy under certain conditions.

Authors

  • Tomoumi Takase
    Graduate School of Information Science and Technology, Hokkaido University, Kita 14 Nishi 9 Kita-ku, Sapporo, Japan. Electronic address: takase_t@complex.ist.hokudai.ac.jp.
  • Satoshi Oyama
    Graduate School of Information Science and Technology, Hokkaido University, Kita 14 Nishi 9 Kita-ku, Sapporo, Japan. Electronic address: oyama@ist.hokudai.ac.jp.
  • Masahito Kurihara
    Graduate School of Information Science and Technology, Hokkaido University, Kita 14 Nishi 9 Kita-ku, Sapporo, Japan. Electronic address: kurihara@ist.hokudai.ac.jp.