Abstract:
This study focuses on identifying texts related to maritime contexts using an advanced Large Language Model (LLM) and cost-sensitive approach for handling data imbalances...Show MoreMetadata
Abstract:
This study focuses on identifying texts related to maritime contexts using an advanced Large Language Model (LLM) and cost-sensitive approach for handling data imbalances. Firstly, a comprehensive dataset specifically for maritime-context queries is collected and augmented. Secondly, the dynamic contextual representations of input query considering the context of each word are obtained by a pre-trained LLM which incorporates Bidirectional Encoder Representations from Transformers (BERT) and Convolutional Neural Network (CNN). Thirdly, a Multi-Layer Perceptron (MLP) is constructed as the classifier to fine-tune the whole network on the newly collected dataset. Finally, the Focal loss is introduced for more effective parameter optimization to tackle the challenge of data imbalance between positive and negative samples, Extensive experiments have been conducted and the following promising results have been obtained: 1) The proposed approach achieves an impressive 99.97% F1 score in recognizing maritime-context texts; 2) The ConvBERT model, an enhancement over the original BERT, demonstrates superior performance in text representation while being more computationally efficient; 3) The Focal loss method outperforms other cost-sensitive learning strategies like class weighting and oversampling techniques; and 4) the proposed method surpasses other deep learning and BERT-based methods in text classification tasks.
Published in: 2024 IEEE Conference on Artificial Intelligence (CAI)
Date of Conference: 25-27 June 2024
Date Added to IEEE Xplore: 30 July 2024
ISBN Information: