Hadoop: Addressing challenges of Big Data
Hadoop: Addressing challenges of Big Data. Hadoop is an open source cloud computing platform of the Apache Foundation that provides a software programming framework called MapReduce and distributed file system, HDFS. It is a Linux based set of tools that uses commodity hardware, which are relatively inexpensive, to handle, analyze and transform large quantity of data.
Hadoop Distributed File System, HDFS, stores huge data set reliably and streams it to user application at high bandwidth and MapReduce is a framework that is used for processing massive data sets in a distributed fashion over a several machines. This paper gives a brief overview of Big Data, Hadoop MapReduce and Hadoop Distributed File System along with its architecture.
Similar IEEE Project Titles
- DataMPI: Extending MPI to Hadoop-Like Big Data Computing
- Research on big data information retrieval based on hadoop architecture.
- Effectiveness Assessment of Solid-State Drive Used in Big Data Services
- Parallel Processing of Big Data Using Power Iteration Clustering over MapReduce.
- Use of Big Data technology in Vehicular Ad-hoc Networks.