TY - GEN
T1 - Hyper-Parameter Tuning based on Random Search for DenseNet Optimization
AU - Nugroho, Ari
AU - Suhartanto, Heru
N1 - Funding Information:
This work is supported by Saintek scholarship given to the first author from Ministry of Research, Technology, and Higher Education, Indonesia.
Publisher Copyright:
© 2020 IEEE.
Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
PY - 2020/9/24
Y1 - 2020/9/24
N2 - Deep learning is a machine learning technology that is currently experiencing rapid development. One of the deep learning architecture models is Densely Connected Convolutional Networks (DenseNet), which connects each layer along with feature maps to all next layers. In other words, the next layer receives input feature maps from all previous layers. Because of this density architecture, the training time is longer and requires much memory. One of the ways to overcome these problems is to optimize using the batching strategy method. This batching method uses adaptive batch size with the ratio of learning rate during training to get faster training time without reducing accuracy. To obtain the optimal performance of this batching, a proper selection of hyper-parameter learning rate and batch size is needed. In this study, we perform tuning this hyper-parameter to get the optimal value. The selection of hyper-parameters uses a random search to select candidates for learning rate and batch size, and then an experiment is conducted on the candidates. Random search randomly selects the hyper-parameters, then training and scoring to the hyper-parameters. Our experiment results show that the lower limit of the batch size used is 64, while the optimal learning rate range is in the range of 0.1 - 0.3. The average accuracy obtained from using this hyper-parameter is 95%. Our experiments run on GPU RTX 2080 Ti with CIFAR-10 and CIFAR-100 datasets.
AB - Deep learning is a machine learning technology that is currently experiencing rapid development. One of the deep learning architecture models is Densely Connected Convolutional Networks (DenseNet), which connects each layer along with feature maps to all next layers. In other words, the next layer receives input feature maps from all previous layers. Because of this density architecture, the training time is longer and requires much memory. One of the ways to overcome these problems is to optimize using the batching strategy method. This batching method uses adaptive batch size with the ratio of learning rate during training to get faster training time without reducing accuracy. To obtain the optimal performance of this batching, a proper selection of hyper-parameter learning rate and batch size is needed. In this study, we perform tuning this hyper-parameter to get the optimal value. The selection of hyper-parameters uses a random search to select candidates for learning rate and batch size, and then an experiment is conducted on the candidates. Random search randomly selects the hyper-parameters, then training and scoring to the hyper-parameters. Our experiment results show that the lower limit of the batch size used is 64, while the optimal learning rate range is in the range of 0.1 - 0.3. The average accuracy obtained from using this hyper-parameter is 95%. Our experiments run on GPU RTX 2080 Ti with CIFAR-10 and CIFAR-100 datasets.
KW - batching strategy
KW - DenseNet
KW - hyper-parameter
KW - random search
UR - http://www.scopus.com/inward/record.url?scp=85097272943&partnerID=8YFLogxK
U2 - 10.1109/ICITACEE50144.2020.9239164
DO - 10.1109/ICITACEE50144.2020.9239164
M3 - Conference contribution
AN - SCOPUS:85097272943
T3 - 7th International Conference on Information Technology, Computer, and Electrical Engineering, ICITACEE 2020 - Proceedings
SP - 96
EP - 99
BT - 7th International Conference on Information Technology, Computer, and Electrical Engineering, ICITACEE 2020 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 7th International Conference on Information Technology, Computer, and Electrical Engineering, ICITACEE 2020
Y2 - 24 September 2020 through 25 September 2020
ER -