Abstract:
Training the deep learning models involves learning of the parameters to meet the objective function. Typically the objective is to minimize the loss incurred during the ...Show MoreMetadata
Abstract:
Training the deep learning models involves learning of the parameters to meet the objective function. Typically the objective is to minimize the loss incurred during the learning process. In a supervised mode of learning, a model is given the data samples and their respective outcomes. When a model generates an output, it compares it with the desired output and then takes the difference of generated and desired outputs and then attempts to bring the generated output close to the desired output. This is achieved through optimization algorithms. An optimization algorithm goes through several cycles until convergence to improve the accuracy of the model. There are several types of optimization methods developed to address the challenges associated with the learning process. Six of these have been taken up to be examined in this study to gain insights about their intricacies. The methods investigated are stochastic gradient descent, nesterov momentum, rmsprop, adam, adagrad, adadelta. Four datasets have been selected to perform the experiments which are mnist, fashionmnist, cifar10 and cifar100. The optimal training results obtained for mnist is 1.00 with RMSProp and adam at epoch 200, fashionmnist is 1.00 with rmsprop and adam at epoch 400, cifar10 is 1.00 with rmsprop at epoch 200, cifar100 is 1.00 with adam at epoch 100. The highest testing results are achieved with adam for mnist, fashionmnist, cifar10 and cifar100 are 0.9826, 0.9853, 0.9855, 0.9842 respectively. The analysis of results shows that adam optimization algorithm performs better than others at testing phase and rmsprop and adam at training phase.
Date of Conference: 10-11 January 2019
Date Added to IEEE Xplore: 16 March 2020
ISBN Information: