To build strong foundation by exploring Hadoop ecosystem with real-world examples.
About This Video
- Get a better understanding of how to set up a HDFS cluster between local storage and the Hadoop filesystem
- Run your own Hadoop clusters on your own machine or in the cloud
- Implement the best practices for Hadoop development
Hadoop emerged in response to the proliferation of masses and masses of data collected by organizations, offering a strong solution to store, process, and analyze what has commonly become known as Big Data. It comprises a comprehensive stack of components designed to enable these tasks on a distributed scale, across multiple servers and thousands of machines.
This course introduces you to the powerful system synonymous with Big Data, demonstrating how to create an instance and leverage Hadoop ecosystem's many components to store, process, manage, and query massive data sets with confidence.
The video course opens with an introduction to the world of Hadoop, where we discuss Nodes, Data Sets, and operations such as map and reduce. The second section deals HDFS, Hadoop's file-system used to store data. Further on, you’ll discover the differences between jobs and tasks, and get to know about the Hadoop UI. After this, we turn our attention to storing data in HDFS and Data Transformations. Lastly, we will learn how to implement an algorithm in Hadoop map-reduce way and analyze the overall performance.