Exploring the applications and methods of processing and storing massive data in the Cloudy environment

Payman kalani torbaghan1 maryam kheirabadi2 Reza Ghaemi3

1) Department of Computer Engineering, Neyshabour Branch, Islamic Azad University, Neyshabour, Iran
2) Department of Computer Engineering ,Neyshabour Branch, Islamic Azad University, Neyshabour , Iran
3) Department of Computer Engineering, Quchan Branch, Islamic Azad University, Quchan,iran

Publication : 3.rd International Congress of Science, Engineering and Technology - Hamburg(germanconf.com)
Abstract :
Today, with the growing number of tools such as social networking and the emergence of concepts such as the semantic web, the volume of data and processing in the large hub system has grown tremendously. For example, a search engine prepares user searches in a fraction of a second resulting from an efficient analysis of massive information collected from the web. Therefore, a mechanism for processing massive information at an affordable cost is very important. Also one of the most widely used cloud computing aspects is the processing of large data sets. The Hadoop Open Source Framework is a cloud-based storage for processing and processing of this type of massive data provided by Apache, which is also more economical due to its open source. Our aim in this study is to examine the advantages and disadvantages of two important parts of the Hadoop distributed file system designed for massive data management and the reducer mapping service that provides a framework for processing massive data in a distributed environment.
Keywords : Cloud Computing massive Data Hadoop MapReduce HDFS