We all use smartphones but have you ever wondered how much data it generates in the form of texts phone calls emails photos videos searches and music approximately 40 exabytes of data gets generated every month by a single smartphone user now imagine this number multiplied by 5 billion smartphone users that’s a lot for our mind to even process isn’t it in fact this amount of data is quite a lot for traditional computing systems to handle and this massive amount of data is what we term as big data let’s have a look at the data generated per minute on the internet 2 point 1 million snaps are shared on snapchat 3.8 million search queries are made on Google.
1 million people log on to Facebook 4 point 5 million videos are watched on YouTube 188 million emails are sent that’s a lot of data so how do you classify any data as big data this is possible with the concept of 5 V’s volume velocity variety veracity and value let us understand this with an example from the healthcare industry hospitals and clinics across the world generate massive volumes of data 2314 exabytes of data are collected annually in the form of patient records and test results all this data is generated at a very high speed which attributes to the velocity of big data.
Variety refers to the various data types such as structured semi-structured and unstructured data examples include Excel records log files and x-ray images accuracy and trustworthiness of the generated data is termed as veracity. Analyzing all this data will benefit the medical sector by enabling faster disease detection better treatment and reduced cost this is known as the value of Big Data but how do we store and process this big day to do this job we have various framework such as Cassandra Hadoop and spark?
Let us take Hadoop as an example and see how Hadoop stores and processes. Big Data Hadoop uses a distributed file system known as Hadoop distributed file system to store big data if you have a huge file your file will be broken down into smaller chunks and stored in various machines not only that when you break the file you also make copies of it which goes into different nodes this way you store your big data in a distributed way and make sure that even if one machine fails your data is safe on another MapReduce technique is used to process big data a lengthy task a is broken into smaller tasks B C and D.
Now instead of one machine three machines take up each task and complete it in a parallel fashion and assemble the results at the end thanks to this the processing becomes easy and fast this is known as parallel processing now that we have stored and processed our big data we can analyze this data for numerous applications in games like Halo 3 and Call of Duty designers analyze user data to understand at which stage most of the users pause restart or quit playing this insight can help them rework on the storyline of the game and improve the user experience which in turn reduces the customer churn rate.
Similarly, Big Data also helped with disaster management during Hurricane sandy in 2012 it was used to gain a better understanding of the effect of the storm on the east coast of the US and necessary measures were taken. It could predict the Hurricanes landfall five days in advance which wasn’t possible earlier these are some of the clear indications of how valuable big data can be once it is accurately processed and analyzed so here’s a question for you which of the following statements is not correct about Hadoop distributed file system HDFS is the storage layer of Hadoop B data gets stored in a distributed manner in HDFS C HDFS performs parallel processing of data D smaller chunks of data are stored on multiple data nodes in HDFS give it a thought.