z-logo
open-access-imgOpen Access
Streaming chunk incremental learning for class-wise data stream classification with fast learning speed and low structural complexity
Author(s) -
Prem Junsawang,
Suphakant Phimoltares,
Chidchanok Lursinsap
Publication year - 2019
Publication title -
plos one
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.99
H-Index - 332
ISSN - 1932-6203
DOI - 10.1371/journal.pone.0220624
Subject(s) - computer science , set (abstract data type) , geodetic datum , class (philosophy) , data stream , artificial intelligence , interval (graph theory) , synthetic data , function (biology) , process (computing) , artificial neural network , streaming algorithm , data set , streaming data , machine learning , data mining , mathematics , telecommunications , mathematical analysis , cartography , combinatorics , evolutionary biology , upper and lower bounds , biology , programming language , geography , operating system
Due to the fast speed of data generation and collection from advanced equipment, the amount of data obviously overflows the limit of available memory space and causes difficulties achieving high learning accuracy. Several methods based on discard-after-learn concept have been proposed. Some methods were designed to cope with a single incoming datum but some were designed for a chunk of incoming data. Although the results of these approaches are rather impressive, most of them are based on temporally adding more neurons to learn new incoming data without any neuron merging process which can obviously increase the computational time and space complexities. Only online versatile elliptic basis function (VEBF) introduced neuron merging to reduce the space-time complexity of learning only a single incoming datum. This paper proposed a method for further enhancing the capability of discard-after-learn concept for streaming data-chunk environment in terms of low computational time and neural space complexities. A set of recursive functions for computing the relevant parameters of a new neuron, based on statistical confidence interval, was introduced. The newly proposed method, named streaming chunk incremental learning (SCIL), increases the plasticity and the adaptabilty of the network structure according to the distribution of incoming data and their classes. When being compared to the others in incremental-like manner, based on 11 benchmarked data sets of 150 to 581,012 samples with attributes ranging from 4 to 1,558 formed as streaming data, the proposed SCIL gave better accuracy and time in most data sets.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here