A Novel Learning Algorithm to Optimize Deep Neural Networks: Evolved Gradient Direction Optimizer (EVGO).

Journal: IEEE transactions on neural networks and learning systems
Published Date:

Abstract

Gradient-based algorithms have been widely used in optimizing parameters of deep neural networks' (DNNs) architectures. However, the vanishing gradient remains as one of the common issues in the parameter optimization of such networks. To cope with the vanishing gradient problem, in this article, we propose a novel algorithm, evolved gradient direction optimizer (EVGO), updating the weights of DNNs based on the first-order gradient and a novel hyperplane we introduce. We compare the EVGO algorithm with other gradient-based algorithms, such as gradient descent, RMSProp, Adagrad, momentum, and Adam on the well-known Modified National Institute of Standards and Technology (MNIST) data set for handwritten digit recognition by implementing deep convolutional neural networks. Furthermore, we present empirical evaluations of EVGO on the CIFAR-10 and CIFAR-100 data sets by using the well-known AlexNet and ResNet architectures. Finally, we implement an empirical analysis for EVGO and other algorithms to investigate the behavior of the loss functions. The results show that EVGO outperforms all the algorithms in comparison for all experiments. We conclude that EVGO can be used effectively in the optimization of DNNs, and also, the proposed hyperplane may provide a basis for future optimization algorithms.

Authors

  • Ibrahim Karabayir
    Center for Health Outcomes and Informatics Research, Health Sciences Division, Loyola University Chicago, Maywood, IL; Department of Health Informatics and Data Science, Loyola University Chicago, Chicago, IL; Kirklareli University, Kirklareli, Turkey.
  • Oguz Akbilgic
    1Department of Pediatrics, University of Tennessee Health Science Center - Oak Ridge National Laboratory- (UTHSC-ORNL), Center for Biomedical Informatics, Memphis, TN USA.
  • Nihat Tas