Impact Statement:Time series data has been widely applied to various real-world applications, e.g., sleep staging, electromyography signal classification, and arrhythmic heartbeat classif...Show More
Abstract:
Over the years, many deep learning algorithms have been developed for time series classification (TSC). A learning model's performance usually depends on the quality of t...Show MoreMetadata
Impact Statement:
Time series data has been widely applied to various real-world applications, e.g., sleep staging, electromyography signal classification, and arrhythmic heartbeat classification. The performance of a learning model relies on the representations captured from the data for download tasks, such as classification. In this article, we introduce self-BiDecKD for TSC. Unlike most self-distillation algorithms that usually transfer the target-class knowledge from higher to lower levels, self-BiDecKD encourages the output of the output layer and the output of each lower level block to form a bidirectional decoupled KD pair. When compared to several state-of-the-art self-distillation algorithms, self-BiDecKD consistently demonstrates outstanding performance across a wide range of TSC applications.
Abstract:
Over the years, many deep learning algorithms have been developed for time series classification (TSC). A learning model's performance usually depends on the quality of the semantic information extracted from lower and higher levels within the representation hierarchy. Efficiently promoting mutual learning between higher and lower levels is vital to enhance the model's performance during model learning. To this end, we propose a self-bidirectional decoupled distillation (self-BiDecKD) method for TSC. Unlike most self-distillation algorithms that usually transfer the target-class knowledge from higher to lower levels, self-BiDecKD encourages the output of the output layer and the output of each lower level block to form a bidirectional decoupled knowledge distillation (KD) pair. The bidirectional decoupled KD promotes mutual learning between lower and higher level semantic information and extracts the knowledge hidden in the target and nontarget classes, helping self-BiDecKD capture ric...
Published in: IEEE Transactions on Artificial Intelligence ( Volume: 5, Issue: 8, August 2024)