This study proposes MDCNN, a Multi-teacher Distillation-based CNN for news text classification. By leveraging BERT and BiGRU as teacher models and TextCNN as the student ...
Abstract:
News text classification is crucial for efficient information acquisition and dissemination. While deep learning models, such as BERT and BiGRU, excel in accuracy for tex...Show MoreMetadata
Abstract:
News text classification is crucial for efficient information acquisition and dissemination. While deep learning models, such as BERT and BiGRU, excel in accuracy for text classification, their high complexity and resource demands hinder practical deployment. To address these challenges, we proposed MDCNN (Multi-teacher Distillation-based CNN), which leverages knowledge distillation with BERT and BiGRU as teacher models and TextCNN as the student model. Experiments on three benchmark news dataset demonstrate that MDCNN improves classification accuracy by nearly 2% while significantly reducing computational overhead, offering a practical solution for real-world applications.
This study proposes MDCNN, a Multi-teacher Distillation-based CNN for news text classification. By leveraging BERT and BiGRU as teacher models and TextCNN as the student ...
Published in: IEEE Access ( Volume: 13)