논문 상세보기

Performance Analysis of Different Optimizers, Batch Sizes, and Epochs on Convolutional Neural Network for Image Classification KCI 등재

  • 언어ENG
  • URLhttps://db.koreascholar.com/Article/Detail/406667
구독 기관 인증 시 무료 이용이 가능합니다. 4,000원
농업생명과학연구 (Journal of Agriculture & Life Science)
경상대학교 농업생명과학연구원 (Institute of Agriculture & Life Science, Gyeongsang National University)
초록

The important thing in the field of deep learning is to find out the appropriate hyper-parameter for image classification. In this study, the main objective is to investigate the performance of various hyper-parameters in a convolutional neural network model based on the image classification problem. The dataset was obtained from the Kaggle dataset. The experiment was conducted through different hyper-parameters. For this proposal, Stochastic Gradient Descent without momentum (SGD), Adaptive Moment Estimation (Adam), Adagrad, Adamax optimizer, and the number of batch sizes (16, 32, 64, 120), and the number of epochs (50, 100, 150) were considered as hyper-parameters to determine the losses and accuracy of a model. In addition, Binary Cross-entropy Loss Function (BCLF) was used for evaluating the performance of a model. In this study, the VGG16 convolutional neural network was used for image classification. Empirical results demonstrated that a model had minimum losses obtain by Adagrad optimizer in the case of 16 batch sizes and 50 epochs. In addition, the SGD with a 32 batch sizes and 150 epochs and the Adam with a 64 batch sizes and 50 epochs had the best performance based on the loss value during the training process. Interestingly, the accuracy was higher while performing the Adagrad and Adamax optimizer with a 120 batch sizes and 150 epochs. In this study, the Adagrad optimizer with a 120 batch sizes and 150 epochs performed slightly better among those optimizers. In addition, an increasing number of epochs can improve the performance of accuracy. It can help to create a broader scope for further experiments on several datasets to perceive the suitable hyper-parameters for the convolutional neural network. Dataset: https://www.kaggle.com/c/dogs-vs-cats/data

목차
Abstract
Introduction
Materials and Methods
    1. Experiment
    2. Batch Size and Epoch
    3. Loss function evaluation criteria
Results and Discussion
    1. Loss evaluation
    2. Accuracy evaluation
    3. Performance of increasing batch size and poch
References
저자
  • Thavisack Sihalath(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm))
  • Jayanta Kumar Basak(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm))
  • Anil Bhujel(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm))
  • Elanchezhian Arulmozhi(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm))
  • Byeong-Eun Moon(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm))
  • Na-Eun Kim(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm))
  • Doeg-Hyun Lee(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm))
  • Hyeon-Tae Kim(Department of Bio-systems Engineering, Gyeongsang National University (Institute of Smart Farm)) Corresponding author