z-logo
open-access-imgOpen Access
Model optimisation of class imbalanced learning using ensemble classifier on over-sampling data
Author(s) -
Yulia Ery Kurniawati,
Yulius Denny Prabowo
Publication year - 2022
Publication title -
iaes international journal of artificial intelligence
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.341
H-Index - 7
eISSN - 2252-8938
pISSN - 2089-4872
DOI - 10.11591/ijai.v11.i1.pp276-283
Subject(s) - computer science , oversampling , adaboost , machine learning , artificial intelligence , classifier (uml) , data mining , ensemble learning , boosting (machine learning) , bandwidth (computing) , computer network
Data imbalance is one of the problems in the application of machine learning and data mining. Often this data imbalance occurs in the most essential and needed case entities. Two approaches to overcome this problem are the data level approach and the algorithm approach. This study aims to get the best model using the pap smear dataset that combined data levels with an algorithmic approach to solve data imbalanced. The laboratory data mostly have few data and imbalance. Almost in every case, the minor entities are the most important and needed. Over-sampling as a data level approach used in this study is the synthetic minority oversampling technique-nominal (SMOTE-N) and adaptive synthetic-nominal (ADASYN-N) algorithms. The algorithm approach used in this study is the ensemble classifier using AdaBoost and bagging with the classification and regression tree (CART) as learner-based. The best model obtained from the experimental results in accuracy, precision, recall, and f-measure using ADASYN-N and AdaBoost-CART.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here