z-logo
open-access-imgOpen Access
Integration of synthetic minority oversampling technique for imbalanced class
Author(s) -
Noviyanti Santoso,
Wahyu Wibowo,
Hilda Hikmawati
Publication year - 2019
Publication title -
indonesian journal of electrical engineering and computer science
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.241
H-Index - 17
eISSN - 2502-4760
pISSN - 2502-4752
DOI - 10.11591/ijeecs.v13.i1.pp102-108
Subject(s) - oversampling , undersampling , support vector machine , class (philosophy) , naive bayes classifier , computer science , random forest , machine learning , artificial intelligence , data mining , statistical classification , pattern recognition (psychology) , bandwidth (computing) , computer network
In the data mining, a class imbalance is a problematic issue to look for the solutions. It probably because machine learning is constructed by using algorithms with assuming the number of instances in each balanced class, so when using a class imbalance, it is possible that the prediction results are not appropriate. They are solutions offered to solve class imbalance issues, including oversampling, undersampling, and synthetic minority oversampling technique (SMOTE). Both oversampling and undersampling have its disadvantages, so SMOTE is an alternative to overcome it. By integrating SMOTE in the data mining classification method such as Naive Bayes, Support Vector Machine (SVM), and Random Forest (RF) is expected to improve the performance of accuracy. In this research, it was found that the data of SMOTE gave better accuracy than the original data. In addition to the three classification methods used, RF gives the highest average AUC, F-measure, and G-means score.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here