z-logo
open-access-imgOpen Access
TAXONOMY OF DATA INCONSISTENCIES IN BIG DATA
Author(s) -
Vinaya Keskar,
Jyoti Yadav,
Ajay Kumar
Publication year - 2021
Publication title -
information technology in industry/information technology in industry
Language(s) - English
Resource type - Journals
eISSN - 2204-0595
pISSN - 2203-1731
DOI - 10.17762/itii.v9i1.139
Subject(s) - terabyte , data quality , data science , computer science , big data , quality (philosophy) , measure (data warehouse) , data governance , data mining , information retrieval , engineering , operating system , metric (unit) , philosophy , epistemology , operations management
In the coming years, common units of measuring data viz. kilobytes, megabytes, gigabytes, or even terabytes will begin to appear quainter as the entire digital universe is expected to produce approximately 463 Exabyte’s of data every 24 hours worldwide. This omnipresent data is potentially knowledge-rich. Unprocessed data can be excavated for hidden information. Essentially, the quality of the output depends on the quality of input data. Alternately, a good analysis of faulty/bad data cannot result in meaningful outputs. The global challenge that arises during data analysis is the quality of data. Data quality is not intentionally reduced by unscrupulous systemic elements as inconsistencies have an uncanny way of creeping in due to various factors. The importance of data allows organizations to measure past performance quantitatively as well as to quantitatively ascertain present capabilities and thereby plan for future performance targets, leading to the study of data inconsistencies. This paper presents a conceptual outline of various categories and types of data inconsistencies, extending it further to briefly explain the data processing life cycle and the sources of data inconsistencies.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here