Self-Bidirectional Decoupled Distillation for Time Series Classification | IEEE Journals & Magazine | IEEE Xplore

Self-Bidirectional Decoupled Distillation for Time Series Classification


Impact Statement:Time series data has been widely applied to various real-world applications, e.g., sleep staging, electromyography signal classification, and arrhythmic heartbeat classif...Show More

Abstract:

Over the years, many deep learning algorithms have been developed for time series classification (TSC). A learning model's performance usually depends on the quality of t...Show More
Impact Statement:
Time series data has been widely applied to various real-world applications, e.g., sleep staging, electromyography signal classification, and arrhythmic heartbeat classification. The performance of a learning model relies on the representations captured from the data for download tasks, such as classification. In this article, we introduce self-BiDecKD for TSC. Unlike most self-distillation algorithms that usually transfer the target-class knowledge from higher to lower levels, self-BiDecKD encourages the output of the output layer and the output of each lower level block to form a bidirectional decoupled KD pair. When compared to several state-of-the-art self-distillation algorithms, self-BiDecKD consistently demonstrates outstanding performance across a wide range of TSC applications.

Abstract:

Over the years, many deep learning algorithms have been developed for time series classification (TSC). A learning model's performance usually depends on the quality of the semantic information extracted from lower and higher levels within the representation hierarchy. Efficiently promoting mutual learning between higher and lower levels is vital to enhance the model's performance during model learning. To this end, we propose a self-bidirectional decoupled distillation (self-BiDecKD) method for TSC. Unlike most self-distillation algorithms that usually transfer the target-class knowledge from higher to lower levels, self-BiDecKD encourages the output of the output layer and the output of each lower level block to form a bidirectional decoupled knowledge distillation (KD) pair. The bidirectional decoupled KD promotes mutual learning between lower and higher level semantic information and extracts the knowledge hidden in the target and nontarget classes, helping self-BiDecKD capture ric...
Published in: IEEE Transactions on Artificial Intelligence ( Volume: 5, Issue: 8, August 2024)
Page(s): 4101 - 4110
Date of Publication: 30 January 2024
Electronic ISSN: 2691-4581

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.