Learning Hadoop 2
4.0 (13 ratings)
Instead of using a simple lifetime average, Udemy calculates a course's star rating by considering a number of different factors such as the number of ratings, the age of ratings, and the likelihood of fraudulent ratings.
122 students enrolled
Wishlisted Wishlist

Please confirm that you want to add Learning Hadoop 2 to your Wishlist.

Add to Wishlist

Learning Hadoop 2

An introduction to storing, structuring, and analyzing data at scale with Hadoop
4.0 (13 ratings)
Instead of using a simple lifetime average, Udemy calculates a course's star rating by considering a number of different factors such as the number of ratings, the age of ratings, and the likelihood of fraudulent ratings.
122 students enrolled
Created by Packt Publishing
Last updated 12/2015
Current price: $10 Original price: $75 Discount: 87% off
1 day left at this price!
30-Day Money-Back Guarantee
  • 1.5 hours on-demand video
  • 1 Supplemental Resource
  • Full lifetime access
  • Access on mobile and TV
  • Certificate of Completion
What Will I Learn?
  • Install and configure an Hadoop instance of your own
  • Navigate Hue, the GUI for common tasks in Hadoop
  • Import data manually, and automatically from a database
  • Build scripts with Pig to perform common ETL tasks
  • Write and run a simple MapReduce program
  • Structure and query data effectively with Hive, Hadoop’s built-in data warehousing component
View Curriculum
  • We expect familiarity working at the Linux command line, and a basic understanding of Java. No prior experience with Hadoop is required.

Hadoop emerged in response to the proliferation of masses and masses of data collected by organizations, offering a strong solution to store, process, and analyze what has commonly become known as Big Data. It comprises a comprehensive stack of components designed to enable these tasks on a distributed scale, across multiple servers and thousands of machines.

Learning Hadoop 2 introduces you to the powerful system synonymous with Big Data, demonstrating how to create an instance and leverage Hadoop ecosystem's many components to store, process, manage, and query massive data sets with confidence.

We open this course by providing an overview of the Hadoop component ecosystem, including HDFS, Sqoop, Flume, YARN, MapReduce, Pig, and Hive, before installing and configuring our Hadoop environment. We take a look at Hue, the graphical user interface of Hadoop.

We will then discover HDFS, Hadoop’s file-system used to store data. We will learn how to import and export data, both manually and automatically. Afterward, we turn our attention toward running computations using MapReduce, and get to grips working with Hadoop’s scripting language, Pig. Lastly, we will siphon data from HDFS into Hive, and demonstrate how it can be used to structure and query data sets.

About The Author

Randal Scott King is the Managing Partner of Brilliant Data, a consulting firm specialized in data analytics. In his 16 years of consulting, Scott has amassed an impressive list of clientele from mid-market leaders to Fortune 500 household names. Scott lives just outside Atlanta, GA, with his children.

Who is the target audience?
  • This video course is designed for application and system developers interested in understanding how to manage and analyze large scale data sets with the Hadoop framework.
Students Who Viewed This Course Also Viewed
Curriculum For This Course
Expand All 19 Lectures Collapse All 19 Lectures 01:30:01
The Hadoop Ecosystem
6 Lectures 25:47

This video will offer the overview of the course.

The Course Overview

This video will introduce you to the basic concepts of Hadoop Distributed File System (HDFS) and Yet Another Resource Negotiator (YARN), which are the two core components of Hadoop.

Overview of HDFS and YARN

An introduction to the basic concepts of Sqoop and Flume, two tools for the automation of data import into Hadoop.

Overview of Sqoop and Flume

An introduction to the basic concepts of MapReduce, the computation engine of Hadoop.

Overview of MapReduce

An introduction to the basic concepts of Pig, a scripting language for Hadoop.

Overview of Pig

An introduction to the basic concepts of Hive, Hadoop’s data warehousing solution.

Overview of Hive
Installing and Configuring Hadoop
2 Lectures 08:23

Put a working Hadoop installation on a laptop or server. You will need Hadoop on your laptop or server in order to continue.

Preview 02:59

Exploring the Hue, a GUI for Hadoop, to get familiar with the interface.

Exploring Hue
Data Import and Export
3 Lectures 16:07

This video will cover how to get data into HDFS manually.

Preview 04:33

This video will explain how to get data from databases into HDFS.

Importing from Databases Using Sqoop

This video will cover how to import streaming data using the Flume tool.

Using Flume to Import Streaming Data
Using MapReduce and Pig
4 Lectures 23:11

This video will explore how to build “Word Count” in Eclipse, then save it to a .jar and run it from MapReduce.

Preview 05:55

Coding the same word counting program, but this time in Pig.

Coding "Word Count" in Pig

This video will discuss how to use Pig to perform common Extract, Transform, and Load functions on data.

Performing Common ETL Functions in Pig

This video will explore how to use predefined code called User Defined Functions (UDFs) in Pig scripts.

Using User-defined Functions in Pig
Using Hive
4 Lectures 16:33

Create a database in Hive.

Preview 04:57

This video will cover how to get data into Hive from a database without going to HDFS first.

Importing Data Directly from a Database

Using queries in Hive to find information.

Performing Basic Queries in Hive

A quick summary of what the viewer has learned in the entire course.

Putting It All Together
About the Instructor
Packt Publishing
3.9 Average rating
4,437 Reviews
35,077 Students
348 Courses
Tech Knowledge in Motion

Packt has been committed to developer learning since 2004. A lot has changed in software since then - but Packt has remained responsive to these changes, continuing to look forward at the trends and tools defining the way we work and live. And how to put them to work.

With an extensive library of content - more than 4000 books and video courses -Packt's mission is to help developers stay relevant in a rapidly changing world. From new web frameworks and programming languages, to cutting edge data analytics, and DevOps, Packt takes software professionals in every field to what's important to them now.

From skills that will help you to develop and future proof your career to immediate solutions to every day tech challenges, Packt is a go-to resource to make you a better, smarter developer.

Packt Udemy courses continue this tradition, bringing you comprehensive yet concise video courses straight from the experts.