z-logo
open-access-imgOpen Access
An Experimental Model of the Collect/Report Paradigm
Author(s) -
Krassimira Ivanova
Publication year - 2021
Publication title -
iop conference series. materials science and engineering
Language(s) - English
Resource type - Journals
eISSN - 1757-899X
pISSN - 1757-8981
DOI - 10.1088/1757-899x/1031/1/012073
Subject(s) - computer science , scalability , cloud computing , big data , simplicity , distributed computing , key (lock) , process (computing) , programming paradigm , server , data science , popularity , database , data mining , world wide web , computer security , operating system , programming language , psychology , social psychology , philosophy , epistemology
In the Big Data community, the “Map/Reduce Paradigm” is one of the key enabling approaches for meeting the continuously increasing demands on computing resources imposed by massive data sets. Today it is implemented in many open source projects. The popularity of Map/Reduce is due to its high scalability, fault-tolerance, simplicity and independence from the programming language or the data storage system. At the same time, Map/Reduce faces a number of obstacles when dealing with Big Data. A possible solution of them may be the Collect/Report Paradigm (CRP) and Natural Language Addressing (NLA) approach. It is suitable for storing Big Data in large information bases located on different storage systems – from personal computers up to cloud servers. An experimental Model of the CRP is presented in this paper. An experimental implementation of the CRP to process and store data is outlined. The structures of the input and output data are in the form of RDF triplets. The ease of implementation of this model and the benefits of its use are discussed.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here