High Impact Factor : 4.396 icon | Submit Manuscript Online icon |

Big Data and Its Implementation using Hadoop

Author(s):

Suraj Bala , Global Institute of Technology and Management, Gurugram; Prof. Anil Mishra, GITM, Gurugram

Keywords:

Big Data, Cluster, Data Processing, Hadoop, Extraction, HDFS

Abstract

In this paper, we are showing that how we can store and process high volume of data (terabytes or petabytes) on computer clusters built from commodity (cheap) hardware. For this, we are using “Apache Hadoop” a software framework that supports data-intensive distributed applications under a free license. In this paper a proposed technique for implementing the storage and processing of huge data sets we are installing a Hadoop Vanilla Cluster (Multi node cluster setup - Pseudo Distributed Mode) on Ubuntu Environment. Thereafter, showing how we can do data ingestion, processing and migration of data using different Hadoop Eco – Systems. The success rate of our proposed scheme depends upon the extraction and configuration techniques which have been applied in this work.

Other Details

Paper ID: IJSRDV6I50230
Published in: Volume : 6, Issue : 5
Publication Date: 01/08/2018
Page(s): 366-369

Article Preview

Download Article