The First Lesson About Big Data Aspects
Welcome in the Big Data Classroom with the first lesson about different Big Data aspects. For those who have trouble defining Big data this can be a real eye-opener. In this article I will try to explain some history about the Big Data aspects and an easy explanation of the term.
What is Big Data?
Who else than the most famous research firm in the world of technology; Gartner Inc. can explain it best.
“Big data is high volume, high velocity, and/or high variety information assets that require new forms of processing to enable enhanced decision making, insight discovery and process optimization. By Doug Lany, VP Research at Gartner
You can see that the definition contains already 3 of the 5 V’s. Researchers include to other two V’s in the following years to paint a complete picture about Big Data. Gartner Inc. was the first to complete the term Big Data with the V’s
The 5 V’s of Big Data
The quantity of generated and stored data. The size of the data determines the value and potential insight- and whether it can actually be considered big data or not.
The type and nature of the data. This helps people who analyze it to effectively use the resulting insight.
In this context, the speed at which the data is generated and processed to meet the demands and challenges that lie in the path of growth and development.
Inconsistency of the data set can hamper processes to handle and manage it.
The quality of captured data can vary greatly, affecting accurate analysis.
The use of data to track and control businesses dates back from 7.000 years ago when accounting was introduced in Mesopotamia. The use of this accounting has the goal to record the growth of crops and herds. Accounting principles continued to improve, and in 1663, John Graunt recorded and examined all information about mortality roles in London. He wanted to gain an understanding and build a warning system for the ongoing plague. This makes John Graunt the founding father of statistics.
More recent history of Big Data aspects
In 2005 Roger Mougalas from O’Reilly Media coined the term Big Data for the first time, only a year after they created the term Web 2.0. The tern refers to a large set of data that is almost impossible to manage and process using traditional business intelligence tool.
Nowadays we talk about enormous amounts of data. A couple of years ago it used to be just a couple of megabytes or gigabytes. But do to the immense growth of data and the ability to gain this data from many different big data aspects or big data sources, the data amounts are estimated at 667 exabytes annually in 2014
For this article I would like to express a special thanks to following sources: