Month: February 2019

How MapReduce Organizes Work?

How MapReduce Organizes Work?

Hadoop divides the job into tasks. There are two types of tasks: Map tasks (Splits & Mapping) Reduce tasks (Shuffling, Reducing) as mentioned above. The complete execution process (execution of Map and Reduce tasks, both) is controlled by two types of entities called a Jobtracker: Acts like a master (responsible for complete execution of submitted … Continue reading How MapReduce Organizes Work?

Advertisements
Introduction to Pig, Sqoop, and Hive

Introduction to Pig, Sqoop, and Hive

Apache Pig The Apache Pig is a platform for managing large sets of data which consists of high-level programming to analyze the data. Pig also consists of the infrastructure to evaluate the programs. The advantages of Pig programming is that it can easily handle parallel processes for managing very large amounts of data. The programming … Continue reading Introduction to Pig, Sqoop, and Hive