z-logo
Premium
Random Prism: a noise‐tolerant alternative to Random Forests
Author(s) -
Stahl Frederic,
Bramer Max
Publication year - 2014
Publication title -
expert systems
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.365
H-Index - 38
eISSN - 1468-0394
pISSN - 0266-4720
DOI - 10.1111/exsy.12032
Subject(s) - overfitting , computer science , decision tree , classifier (uml) , random forest , artificial intelligence , ensemble learning , prism , machine learning , decision tree learning , pattern recognition (psychology) , statistical classification , artificial neural network , physics , optics
Ensemble learning can be used to increase the overall classification accuracy of a classifier by generating multiple base classifiers and combining their classification results. A frequently used family of base classifiers for ensemble learning are decision trees. However, alternative approaches can potentially be used, such as the Prism family of algorithms that also induces classification rules. Compared with decision trees, Prism algorithms generate modular classification rules that cannot necessarily be represented in the form of a decision tree. Prism algorithms produce a similar classification accuracy compared with decision trees. However, in some cases, for example, if there is noise in the training and test data, Prism algorithms can outperform decision trees by achieving a higher classification accuracy. However, Prism still tends to overfit on noisy data; hence, ensemble learners have been adopted in this work to reduce the overfitting. This paper describes the development of an ensemble learner using a member of the Prism family as the base classifier to reduce the overfitting of Prism algorithms on noisy datasets. The developed ensemble classifier is compared with a stand‐alone Prism classifier in terms of classification accuracy and resistance to noise.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here