HDP Developer Quick Start

This Course Includes neXT LIVE 365

LEARN FOR 365 DAYS!

Sunset Learning Institute believes in a 365-day learning experience that begins immediately, regardless of when you attend your ILT course.  At SLI, you get a range of learning opportunities, from instructor-led hands-on training, to self-directed, customizable learning paths based on your environment, your needs, and your level of experience. We provide the tools and options, and you decide what you need, when you need it, and how you want to learn it! 

Immediate access to supplemental learning assets that are INCLUDED with your purchase of the above instructor-led training course: 
  • 365 Days of Access to SLI’s Entire Big Data Video Reference Library (VRL), not just the 5-day class you sign up for (hundreds of searchable, on-demand learning bytes in 5-15-minute videos)
  • 365 Days of Unlimited Access to Delta Sessions - What’s Not Covered in Class! (Version Upgrades, Industry Updates, Etc.)
  • 365 Days of Unlimited 24x7 Access to SLI's Community - Collaborate with SLI Instructors and Other Members (Monitored Daily by SLI Instructors) See Community Demo
  • 365 Days of Unlimited Access to Interactive neXTpertise Sessions and other IT Resources with SLI Instructors (featured hot topics, exam prep, etc.)  See Upcoming neXTpertise Sessions
  • Unlimited Access to Hosted Webinars and All Previously Recorded Sessions
  • Unlimited Access to your Digital Courseware
See Entire Portfolio

Benefits:
  • Training that fits your needs (from high intensity to small learning bytes)
  • Build immediate competency - start at time of purchase!
  • Gain know-how and skills gaps with limited work disruptions
  • Get quick answers to daily challenges - live interaction!


Overview

This 4 day training course is designed for developers who need to create applications to analyze Big Data stored in Apache Hadoop using Apache Pig and Apache Hive, and developing applications on Apache Spark.

Topics include: Essential understanding of HDP and its capabilities, Hadoop, YARN, HDFS, MapReduce/Tez, data ingestion, using Pig and Hive to perform data analytics on Big Data and an introduction to Spark Core, Spark SQL, Apache Zeppelin, and additional Spark features.
 

Target Audience

Developers and data engineers who need to understand and develop applications on HDP.

Prerequisites

Students should be familiar with programming principles and have experience in software development. SQL and light scripting knowledge is also helpful. No prior Hadoop knowledge is required.  Target Audience Developers and data engineers who need to understand and develop applications on HDP.

Course Outline

Day 1: An Introduction to Apache Hadoop and HDFS

  • Describe the Case for Hadoop
  • Describe the Trends of Volume, Velocity and Variety
  • Discuss the Importance of Open Enterprise Hadoop
  • Describe the Hadoop Ecosystem Frameworks Across the Following Five Architectural Categories:
    • Data Management
    • Data Access
    • Data Governance & Integration
    • Security
    • Operations
  • Describe the Function and Purpose of the Hadoop Distributed File System (HDFS)
  • List the Major Architectural Components of HDFS and their Interactions
  • Describe Data Ingestion
  • Describe Batch/Bulk Ingestion Options
  • Describe the Streaming Framework Alternatives
  • Describe the Purpose and Function of MapReduce
  • Describe the Purpose and Components of YARN
  • Describe the Major Architectural Components of YARN and their Interactions
  • Define the Purpose and Function of Apache Pig
  • Work with the Grunt Shell
  • Work with Pig Latin Relation Names and Field Names
  • Describe the Pig Data Types and Schema

Day 1 Labs and Demonstrations
  • Starting an HDP Cluster
  • Using HDFS Commands
  • Demonstration: Understanding Apache Pig
  • Getting Started with Apache Pig
  • Exploring Data with Pig

Day 2: Advanced Apache Pig Programming
  • Demonstrate Common Operators Such as:
    • Order by
    • Case
    • Distinct
    • Parallel
    • Foreach
  • Understand how Hive Tables are Defined and Implemented
  • Use Hive to Explore and Analyze Data Sets
  • Explain and Use the Various Hive File Formats
  • Create and Populate a Hive Table that Uses ORC File Formats
  • Use Hive to Run SQL-like Queries to Perform Data Analysis
  • Use Hive to Join Datasets Using a Variety of Techniques
  • Write Efficient Hive Queries
  • Explain the Uses and Purpose of HCatalog
  • Use HCatalog with Pig and Hive

Day 2 Labs and Demonstrations
  • Splitting a Dataset
  • Joining Datasets
  • Preparing Data for Apache Hive
  • Understanding Apache Hive Tables
  • Demonstration: Understanding Partitions and Skew
  • Analyzing Big Data with Apache Hive
  • Demonstration: Computing Ngrams
  • Joining Datasets in Apache Hive
  • Computing Ngrams of Emails in Avro Format
  • Using HCatalog with Apache Pig

Day 3: Advanced Apache Pig Programming
  • Describe How to Perform a Multi-Table/File Insert
  • Define and Use Views
  • Define and Use Clauses and Windows
  • List the Hive File Formats Including:
    • Text Files
    • SequenceFile
    • RCFile
    • ORC File
  • Define Hive Optimization
  • Use Apache Zeppelin to Work with Spark
  • Describe the Purpose and Benefits of Spark
  • Define Spark REPLs and Application Architecture
  • Explain the Purpose and Function of RDDs
  • Explain Spark Programming Basics
  • Define and Use Basic Spark Transformations
  • Define and Use Basic Spark Actions
  • Invoke Functions for Multiple RDDs, Create Named Functions and Use Numeric Operations

Day 3 Labs
  • Advanced Apache Hive Programming
  • Introduction to Apache Spark REPLs and Apache Zeppelin
  • Creating and Manipulating RDDs
  • Creating and Manipulating Pair RDDs

Day 4: Working with Pair RDDS and Building Yarn Applications
  • Define and Create Pair RDDs
  • Perform Common Operations on Pair RDDs
  • Name the Various Components of Spark SQL and Explain their Purpose
  • Describe the Relationship Between DataFrames, Tables and Contexts
  • Use Various Methods to Create and Save DataFrames and Tables
  • Understand Caching, Persisting and the Different Storage Levels
  • Describe and Implement Checkpointing
  • Create an Application to Submit to the Cluster
  • Describe Client vs Cluster Submission with YARN
  • Submit an Application to the Cluster
  • List and Set Important Configuration Items

Day 4 Labs
  • Creating and Saving DateFrames and Tables
  • Working with DataFrames
  • Building and Submitting Applications to YARN

 

SLI Main Menu