Random Sharing Parameters in the Global Region of Convolutional Neural Network | IEEE Journals & Magazine | IEEE Xplore

Random Sharing Parameters in the Global Region of Convolutional Neural Network


Impact Statement:Building a compact structure and less parameters’ model while preserving its competitive performance is always meaningful in the field of deep neural network. In this art...Show More

Abstract:

Building an efficient model with a compact structure and less parameters while preserving its competitive performance is meaningful in the field of neural networks. Tradi...Show More
Impact Statement:
Building a compact structure and less parameters’ model while preserving its competitive performance is always meaningful in the field of deep neural network. In this article, considering the strong learning ability of deep neural models, we replaced the traditional parameter setting way by sharing parameters in the global region of a convolutional neural network model. We applied our method on the standard ResNet and DenseNet models on several benchmark datasets, and the results demonstrate that the new models can still sustain a competitive performance although with a slight decline under some cases. We extend weight sharing from the interior of one feature map to any layers, through which we can control the number of parameters of a neural model based on the requirement.

Abstract:

Building an efficient model with a compact structure and less parameters while preserving its competitive performance is meaningful in the field of neural networks. Traditionally, a unique group of parameters is identified for each convolution layer. Inspired by the universal approximation theorem, in this study, we explore a flexible way to configure the parameters of a convolutional neural network model. First, we set a parameter pool that stores a certain number of parameters, through which we can also control the number of parameters of a neural model. Second, we randomly select a group of continuous position parameters from the pool for each convolution layer. Finally, we perform extensive experiments for the standard architectures of the ResNet and DenseNet on several benchmark datasets. In the experiments, on CIFAR-10, most of the models could perform almost as well as the original ones within a 0.7% decline. On the difficult tasks CIFAR-100 and ImageNet, most of the models perf...
Published in: IEEE Transactions on Artificial Intelligence ( Volume: 3, Issue: 5, October 2022)
Page(s): 738 - 748
Date of Publication: 17 December 2021
Electronic ISSN: 2691-4581

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.