|Course Price: INR 17,995.00|
What is big data and Apache Hadoop. You will also learn how Hadoop solves the big data problems, about Hadoop cluster architecture, its core components & ecosystem, Hadoop data loading & reading mechanism and role of a Hadoop cluster administrator.
Objective:In this module, you will understand what is big data and Apache Hadoop. You will also learn how Hadoop solves the big data problems, about Hadoop cluster architecture, its core components & ecosystem, Hadoop data loading & reading mechanism and role of a Hadoop cluster administrator.
Topics: Introduction to big data, limitations of existing solutions, Hadoop architecture, Hadoop components and ecosystem, data loading & reading from HDFS, replication rules, rack awareness theory, Hadoop cluster administrator: Roles and responsibilities.
Objective: In this module, you will understand different Hadoop components, understand working of HDFS, Hadoop cluster modes, configuration files, and more. You will also understand the Hadoop 1.0 cluster setup and configuration, setting up Hadoop Clients using Hadoop 1.0 and resolve problems simulated from real-time environment.
Topics: Hadoop server roles and their usage, Hadoop installation and initial configuration, deploying Hadoop in a pseudo-distributed mode, deploying a multi-node Hadoop cluster, Installing Hadoop Clients, understanding working of HDFS and resolving simulated problems.
Objective: In this module you will understand the working of the secondary namenode, working with Hadoop distributed cluster, enabling rack awareness, maintenance mode of Hadoop cluster, adding or removing nodes to your cluster in adhoc and recommended way, understand MapReduce programming model in context of Hadoop administrator and schedulers.
Topics:Understanding secondary namenode, working with Hadoop distributed cluster, Decommissioning or commissioning of nodes, understanding MapReduce, understanding schedulers and enabling them.
Objective: In this module, you will understand day to day cluster administration tasks, balancing data in cluster, protecting data by enabling trash, attempting a manual failover, creating backup within or across clusters, safe guarding your metadata and doing metadata recovery or manual failover of NameNode recovery, learn how to restrict the usage of HDFS in terms of count and volume of data, and more.
Topics: Key admin commands like Balancer, Trash, Import Check Point, Distcp, data backup and recovery, enabling trash, namespace count quota or space quota, manual failover or metadata recovery.
Objective: In this module, you will gather insights around cluster planning and management, learn about the various aspects one needs to remember while planning a setup of a new cluster, capacity sizing, understanding recommendations and comparing different distributions of Hadoop, understanding workload and usage patterns and some examples from world of big data.
Topics: Planning a Hadoop 2.0 cluster, cluster sizing, hardware, network and software considerations, popular Hadoop distributions, workload and usage patterns, industry recommendations.
Objective: In this module, you will learn more about new features of Hadoop 2.0, HDFS High Availability, YARN framework and job execution flow, MRv2, federation, limitations of Hadoop 1.x and setting up Hadoop 2.0 Cluster setup in pseudo-distributed and distributed mode.
Topics:Limitations of Hadoop 1.x, features of Hadoop 2.0, YARN framework, MRv2, Hadoop high availability and federation, yarn ecosystem and Hadoop 2.0 Cluster setup.
Objective: In this module, you will learn to setup Hadoop 2 with high availability, upgrading from v1 to v2, importing data from RDBMS into HDFS, understand why Oozie, Hive and Hbase are used and working of the components.
Topics: Configuring Hadoop 2 with high availability, upgrading to Hadoop 2, working with Sqoop, understanding Oozie, working with Hive, working with Hbase.
Objective: In this module, you will learn about Cloudera manager to setup Cluster, optimisations of Hadoop/Hbase/Hive performance parameters and understand basics on Kerberos. You will learn to setup Pig to use in local/distributed mode to perform data analytics.
Topics: Cloudera manager and cluster setup,Hive administration, HBase architecture, HBase setup, Hadoop/Hive/Hbase performance optimization, Pig setup and working with grunt, why Kerberos and how it helps.