Loading [MathJax]/extensions/MathMenu.js
Supervised Local Training With Backward Links for Deep Neural Networks | IEEE Journals & Magazine | IEEE Xplore

Supervised Local Training With Backward Links for Deep Neural Networks


Impact Statement:Deep learning models are generally bulky with abundant parameters, demanding long training time and excessive computing power. It poses a great challenge for deploying th...Show More

Abstract:

The restricted training pattern in the standard backpropagation (BP) requires end-to-end error propagation, causing large memory costs and prohibiting model parallelizati...Show More
Impact Statement:
Deep learning models are generally bulky with abundant parameters, demanding long training time and excessive computing power. It poses a great challenge for deploying these models in real-world applications. This study looks into the intrinsic limitation of the standard training algorithm. The proposed method increases information flow in the algorithm and improves image classification accuracy over conventional methods. It can substantially reduce memory cost and hardware runtime by 50%. With the proposed method, an intelligent system can be readily deployed on mobile devices and function to meet practical demands. It also enables computing systems to reduce excessive energy waste and cut down carbon footprints for deep learning applications, such as traffic control and autonomous driving.

Abstract:

The restricted training pattern in the standard backpropagation (BP) requires end-to-end error propagation, causing large memory costs and prohibiting model parallelization. Existing local training methods aim to resolve the training obstacles by completely cutting off the backward path between modules and isolating their gradients. These methods prevent information exchange between modules and result in inferior performance. This work proposes a novel local training algorithm, BackLink, which introduces intermodule backward dependence and facilitates information to flow backward along with the network. To preserve the computational advantage of local training, BackLink restricts the error propagation length within the module. Extensive experiments performed in various deep convolutional neural networks demonstrate that our method consistently improves the classification performance of local training algorithms over other methods. For example, our method can surpass the conventional gr...
Published in: IEEE Transactions on Artificial Intelligence ( Volume: 5, Issue: 2, February 2024)
Page(s): 854 - 867
Date of Publication: 02 March 2023
Electronic ISSN: 2691-4581

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.