Apache Sqoop for Certifications - CCA and HDPCD

Deep dive into all the nuances of the Sqoop
Rating: 4.4 out of 5 (351 ratings)
8,975 students
Apache Sqoop for Certifications - CCA and HDPCD
Rating: 4.4 out of 5 (351 ratings)
8,975 students
Data ingestion from RDBMS to Hadoop using Apache Sqoop
Prepare Sqoop for CCA 175 Spark and Hadoop Developer
Prepare Sqoop for HDPCD with Hive, Sqoop and Flume

Requirements

  • Basic linux skills
  • Basic programming skills
  • Cloudera Quickstart VM or Hortonworks Sandbox or valid account for IT Versity Big Data labs or any Hadoop clusters where sqoop and hive are integrated.
  • Minimum memory required based on the environment you are using with 64 bit operating system
Description

As part of this course, we will be

  • Seeing various setup options to explore sqoop
  • Understand how to import data from mysql database to Hadoop HDFS/Hive
  • All the important control arguments while performing import
  • Export data from Hive/HDFS to MySQL

After the course, you can confidently execute scenarios related to sqoop as part of certifications and also make better decisions while building data integration frameworks using Sqoop.

Who this course is for:
  • Any IT aspirant/professional willing to learn Sqoop for certifications or projects
Course content
7 sections • 80 lectures • 6h 28m total length
  • Introduction
    06:55
  • Setup Options
    01:48
  • Setup Cloudera QuickStart VM
    07:22
  • Setup Hortonworks Sandbox
    06:16
  • Data Sets and Big Data labs for practicing Sqoop - from ITVersity
    04:43
  • Using Windows - Putty
    10:33
  • Using Windows - Cygwin
    14:46
  • Introduction to Sqoop
    02:46
  • Validate Source Database - MySQL
    05:52
  • Review JDBC Jar file to connect to MySQL
    02:07
  • Getting help of Sqoop using Command Line
    05:26
  • Overview of Sqoop User Guide
    04:31
  • Validate Sqoop and MySQL integration using "sqoop list-databases"
    04:34
  • List tables in MySQL using "sqoop list-tables"
    01:52
  • Run Queries in MySQL using "sqoop eval"
    05:13
  • Understanding Logs in Sqoop
    01:41
  • Redirecting Sqoop Logs into files
    03:21
  • Overview of Sqoop Import Command
    03:40
  • Perform Sqoop Import of orders - --table and --target-dir
    04:36
  • Perform Sqoop import of order_items - --warehouse-dir
    03:32
  • Sqoop Import - Managing HDFS Directories - append or overwrite or fail
    04:46
  • Sqoop Import - Execution Flow
    05:59
  • Reviewing logs of Sqoop Import
    04:31
  • Sqoop Import - Specifying Number of Mappers
    04:28
  • Review the Output Files
    05:32
  • Sqoop Import - Supported File Formats
    03:56
  • Validating avro Files using avro-tools
    04:47
  • Sqoop Import - Using Compression
    06:28
  • Sqoop Import - Customizing - Introduction
    02:47
  • Sqoop Import - Specifying Columns
    04:34
  • Sqoop Import - Using boundary query
    05:27
  • Sqoop Import - Filter unnecessary data
    05:12
  • Sqoop Import - Using Split By
    04:43
  • Sqoop Import - Importing Query Results
    09:12
  • Sqoop Import - Dealing with Composite Keys
    05:11
  • Sqoop Import - Dealing with Primary Key or Split By using Non Numeric Field
    03:56
  • Sqoop Import - Dealing with Tables with out Primary Key
    06:56
  • Sqoop Import - Autoreset to One Mapper
    03:59
  • Sqoop Import - Default Delimiters using Text File Format
    03:12
  • Sqoop Import - Specifying Delimiters - Import NYSE Data with \t as delimiter
    04:21
  • Sqoop Import - Dealing with NULL Values
    04:31
  • Sqoop Import - import-all-tables
    09:53
  • Quick Overview of Hive
    04:07
  • Sqoop Import - Create Hive Database
    01:08
  • Creating empty Hive Table using create-hive-table
    03:39
  • Sqoop Import - Import orders table to Hive Database
    05:11
  • Sqoop Import - Managing Table using Hive Import - Overwrite
    03:09
  • Sqoop Import - Managing Table using Hive Import - Error out - create-hive-table
    04:48
  • Sqoop Import - Understanding Execution Flow while importing into Hive Table
    07:49
  • Sqoop Import - Review files in Hive Tables
    03:17
  • Sqoop Delimiters vs. Hive Delimiters - Text Files
    02:06
  • Sqoop Import - Hive File Formats
    07:54
  • Sqoop Import all tables - Hive
    06:23
  • Introduction
    04:26
  • Prepare data for Export
    06:26
  • Creating Table in MySQL
    07:02
  • Sqoop Export - Perform Simple Export - --table and --export-dir
    04:57
  • Sqoop Export - Execution Flow
    02:50
  • Sqoop Export - Specifying Number of Mappers
    02:12
  • Sqoop Export - Troubleshooting the issues
    02:24
  • Sqoop Export - Merging or Upserting Overview
    02:52
  • Sqoop Export - Quick Overview of MySQL for Upsert
    07:13
  • Sqoop Export - Using update-mode - update-only (default)
    03:46
  • Sqoop Export - Using update-mode - allow-inseert
    07:33
  • Sqoop Export - Specifying Columns
    07:46
  • Sqoop Export - Specifying Delimiters
    06:49
  • Sqoop Export - Using Stage Table
    09:38
  • Overview of Sqoop Jobs
    02:10
  • Adding Password File
    03:52
  • Creating Sqoop Job
    04:00
  • Running Sqoop Job
    04:16
  • Overview of Incremental Imports
    02:41
  • Incremental Import - Using where
    06:24
  • Incremental Import - Append Mode
    02:59
  • Incremental Import - Create training_orders_incr in retail_export
    03:40
  • Incremental Import - Create Job
    03:46
  • Incremental Import - Execute Job
    02:12
  • Incremental Import - Add additional data (order_id > 30000)
    02:12
  • Incremental Import - Rerun the job and validate results
    03:52
  • Incremental Import - Using mode lastmodified
    05:03

Instructors
Technology Adviser and Evangelist
Durga Viswanatha Raju Gadiraju
  • 4.2 Instructor Rating
  • 8,479 Reviews
  • 146,282 Students
  • 19 Courses

13+ years of experience in executing complex projects using vast array of technologies including Big Data and Cloud.

I found itversity, llc - a US based startup to provide quality training for IT professionals and staffing as well as consulting solutions for enterprise clients. I have trained thousands of IT professionals in vast array of technologies including Big Data and Cloud.

Building IT career for people and provide quality services to the clients will be paramount to our organization.

As an entry strategy itversity will be providing quality training in the areas of ABCD

* Application Development
* Big Data and Business Intelligence
* Cloud
* Datawarehousing, Databases

Support Account for ITVersity Courses.
Itversity Support
  • 4.2 Instructor Rating
  • 8,479 Reviews
  • 145,155 Students
  • 19 Courses

We have built a team to support going forward. If you send messages to this account for our courses, they will be sent to our Helpdesk from where we will be rewriting to our team.