CDH5 - Setup Hadoop Core Components on GCP
Provision Servers from GCP and Setup Hadoop Core Components using Cloudera Distribution of Hadoop
Cloudera Distribution is the most popular Big Data
Provision Servers from GCP and Setup Hadoop Core Components using Cloudera Distribution of Hadoop
Cloudera Distribution is the most popular Big Data distribution in the market. As part of this course you will learn how to set up Multinode Big Data cluster using Cloudera Distribution on GCP.
Here is the agenda for the course.
Provision servers from GCP
Setup Ansible and use it to simplify mundane and repetitive tasks.
Install Cloudera Manager on the designated server
Setup Cloudera Manager and Cloudera Distribution of Hadoop on all nodes in the cluster
Capacity Planning to determine the mapping between Services and Hosts
Configure HDFS, YARN+MR2, Hive, Sqoop, Pig, Oozie and Hue
Validate each of the services that are setup on the cluster
Troubleshooting Tips and Techniques
Day to Day Operations on the cluster.
As part of this course we will primarily focus on core services that comes as part of Cloudera Distribution. We will have courses created very soon on topics such as adding additional tools such as Kafka, Spark etc, Securitizing the Cluster in near future.
Who this course is for:
System Administrators who want to quickly learn setting up Big Data Clusters
Level 1/2 support operations team members who want to learn setting up Big Data Clusters
Data Engineers who want to understand how clusters are typically setup
Architects to understand how to setup Big Data clusters for Proof of Concepts
Comments