Abstract:
This study looks at the effect of unbalanced datasets on machine learning models and assesses the effectiveness of the Synthetic Minority Over-sampling Technique (SMOTE) ...Show MoreMetadata
Abstract:
This study looks at the effect of unbalanced datasets on machine learning models and assesses the effectiveness of the Synthetic Minority Over-sampling Technique (SMOTE) in resolving this issue. The study demonstrates early variances in performance indicators using Naive Bayes, Decision Tree, and Logistic Regression models. SMOTE results in significant gains, notably in accuracy for Naive Bayes and Decision Tree, and precision for Logistic Regression. The findings provide nuanced impact of customized oversampling and the balancing strategies to overcome the drawbacks and threats of imbalanced datasets, leading to optimal model performance.
Date of Conference: 26-28 February 2024
Date Added to IEEE Xplore: 22 May 2024
ISBN Information: