Big Data and Its Implementation using Hadoop |
Author(s): |
Suraj Bala , Global Institute of Technology and Management, Gurugram; Prof. Anil Mishra, GITM, Gurugram |
Keywords: |
Big Data, Cluster, Data Processing, Hadoop, Extraction, HDFS |
Abstract |
In this paper, we are showing that how we can store and process high volume of data (terabytes or petabytes) on computer clusters built from commodity (cheap) hardware. For this, we are using “Apache Hadoop†a software framework that supports data-intensive distributed applications under a free license. In this paper a proposed technique for implementing the storage and processing of huge data sets we are installing a Hadoop Vanilla Cluster (Multi node cluster setup - Pseudo Distributed Mode) on Ubuntu Environment. Thereafter, showing how we can do data ingestion, processing and migration of data using different Hadoop Eco – Systems. The success rate of our proposed scheme depends upon the extraction and configuration techniques which have been applied in this work. |
Other Details |
Paper ID: IJSRDV6I50230 Published in: Volume : 6, Issue : 5 Publication Date: 01/08/2018 Page(s): 366-369 |
Article Preview |
|
|