Talend For Big Data Integration Course : Beginner to Expert

Master guide for using Talend Big Data
3.5 (3 ratings) Instead of using a simple lifetime average, Udemy calculates a
course's star rating by considering a number of different factors
such as the number of ratings, the age of ratings, and the
likelihood of fraudulent ratings.
23 students enrolled
$100
Take This Course
  • Lectures 103
  • Length 11.5 hours
  • Skill Level All Levels
  • Languages English
  • Includes Lifetime access
    30 day money back guarantee!
    Available on iOS and Android
    Certificate of Completion
Wishlisted Wishlist

How taking a course works

Discover

Find online courses made by experts from around the world.

Learn

Take your courses with you and learn anywhere, anytime.

Master

Learn and practice real-world skills and achieve your goals.

About This Course

Published 10/2016 English

Course Description

Course Description

Talend Open Studio for Data Integration is an open Source ETL Tool, which means small companies or businesses can use this tool to perform Extract Transform and Load their data into Databases or any File Format (Talend supports many file formats and Database vendors).

Talend Open Studio for Big Data is an open Source Tool used to interact with Big Data systems from Talend.

If you want to learn how to use Talend Open Studio for Big Data from SCRATCH or If you want to IMPROVE your skills in Big Data Concepts and designing Talend Jobs, then this course is right for you.

Its got EVERYTHING, covers almost all the topics in Talend Open Studio for Big Data.

Talks about Real Time USE CASES.

Prepares you for the Certification Exam.

By the end of the Course you will Master Working with Big Data by designing Talend Jobs.

And what more you ask, All the Videos are HD Quality.

What Are the System Requirements ?

  • PC or Mac.
  • Virtual Box Which is FREE.
  • Talend Software Which is FREE.
  • HDP VM Which is FREE.
  • CDH VM Which is FREE.

What are the requirements?

  • Talend Data Integration Basics

What am I going to get from this course?

  • Learn Basic concepts of Big Data (Hadoop)
  • Create cluster Metadata manually, from configuration files and automatically
  • Create HDFS and Hive metadata Connect to your cluster to use HDFS, HBase, Hive, Pig, Sqoop and Map Reduce
  • Read and Write data to/from HDFS (HDFS, HBase)
  • Read and Write tables to/from HDFS (Hive, Sqoop)
  • Processing Tables stored on HDFS with Hive
  • Processing data stored on HDFS with Pig
  • Use Talend Open Studio for Big Data for real work as quickly as possible.
  • Write Talend Big Data v6 Certified Developer Exam
  • Work on Cloudera Hadoop Distribution
  • Work on HortonWorks Hadoop Distribution
  • Over 100 Lectures and Hours of Content !
  • Over 50 Exercises and Quiz Questions!
  • Once you finish this Course I guarantee, you will Pass the Certification Exam. (Offcourse you have to practice what ever I teach in this course :-)).
  • You will get Source code and Data Used in all 50 + Exercises.
  • You will get Source code and Data Used in all 100 + Jobs Designed in the Course.
  • I will respond to all your questions within 24 hours.
  • 40 % Off on my other course (Talend Data Integration Course : Beginner to Expert)

What is the target audience?

  • Any person who wants to use Talend Studio to interact with Big Data systems.

What you get with this course?

Not for you? No problem.
30 day money back guarantee.

Forever yours.
Lifetime access.

Learn on the go.
Desktop, iOS and Android.

Get rewarded.
Certificate of completion.

Curriculum

Section 1: What Does the Course Cover ?
04:25

This Video will talk about the topics that are covered in this course.

04:14

This video will show you how to download data files and Talend jobs that are designed as part of this course.

Section 2: TALEND OVERVIEW
07:10

This lecture gives you an overview of what Talend Open Studio for Big data is and it also talks about the additional features in the subscription version.

03:52

After watching this lecture, you should be able to download and open Talend Open Studio for Big Data on your Windows OS.

Section 3: BIG DATA OVERVIEW
05:21

You should be able to describe the 3 V's of Big Data and what big data is ?

10:42

Introduction to Hadoop and its advantages over traditional systems.

05:21

Walks you through different Hadoop ecosystem Tools.

05:33

What is a HDFS Block, What is a Namenode and its functions, Waht is a DataNode and its functions.

05:40

This lecture explains you, What happens when you Read/Write a file from/to to HDFS.

12:38

Gives an overview of MapReduce and its functions.

09:46

Explains the Map phase, the Shuffle Sort Phase and the Reduce Phase in a Map Reduce Job.

02:22

Explains different types of Key-Values pairs generated as part of a MapReduce Job.

HDFS - HDFS Federation & NameNode High Availability Hadoop 2
04:48
04:03

Briefly explains YARN and its daemons.

03:12

Explains what happens when you run an application on YARN.

Section 4: Getting Started
08:13

After watching this video you should be able to download CDH virtual box image and your cluster should be up and running.

11:57

After watching this video you should be able to download HDP virtual box image and your cluster should be up and running.

04:40

You should be able to create and open a Talend project.

15:20

This lecture walks you through the process of creating a Hadoop cluster Metadata using HDP.

Creating Hadoop Cluster Metadata in Talend for CDH
03:49
Section 5: HDFS Components
17:15

This video will explain what HDFS is and shows you how to run different hdfs commands on cluster.

For example, how to create a file/directory,how to copy file from local system to hdfs, etc..

06:37

It will show you how to create a HDFS connection in a way that you can use it in all of your HDFS jobs.

08:51

Explains the scenario where you have to copy a file or directory from your local system to HDFS.

How to retrieve a list of files or folders based on a filemask pattern ?
06:36
How to copy files from HDFS to HDFS ?
05:11
How to get files from HDFS into local directory ?
07:40
How to rename the selected files or specified directory on HDFS ?
01:52
How to check whether a file exists in a specific directory in HDFS ?
08:34
How to delete a file located on a given HDFS ?
07:30
How to read a file located on a given HDFS and Assign schema to it?
11:02
How to count the number of rows in a file in HDFS ?
03:32
How to present the properties of a file processed in HDFS ?
03:36
How to transfer data flows into a given HDFS file system ?
04:56
How to transfer data in the form of a single column into a given HDFS ?
02:45
10:49

Explains the scenario where you have to compare two different files that are on HDFS.

Section 6: HIVE Components
What is Hive ?
05:06
Hive Architecture
04:45
HiveQL Vs SQL
05:21
How to Connect to Hive Shell
04:06
How to Create Hive Managed and External Tables Using Hive Shell
13:53
How to Load data from HDFS & Local File System to Hive table using Hive Shell
15:12
How to Load data from one Hive table to another Hive table using Hive Shell
Preview
07:06
How to join two HIVE Tables using Hive Shell
10:08
How to READ data from a HIVE Table and filter data using Hive Shell
01:54
How to open a connection to a Hive database using Talend?
25:59
How to close connection to a Hive databases using Talend?
06:13
How to create a Hive table using Talend?
17:26
How to extract data from Hive using Talend?
05:56
How to write data of different formats into a given Hive table using Talend?
12:21
How to execute the HiveQL query using Talend?
06:16
Section 7: PIG Components
What is Pig ?
11:59
What are the different Datatypes supported by Pig ?
06:44
How to Assign a schema to input file using Grunt Shell ?
20:51
What are aliases,relations and How to Load a file into Pig Alias ?
12:20
Pig - GROUP,GROUP ALL,DUMP,STORE,FILTER,LIMIT Operators
25:46
Pig - FOREACH, COUNT, MAX Operators
19:40
Pig - ORDER BY,DISTINCT,JOIN,COGROUP
21:06
How to load input data to an output stream in one single transaction ?
19:10
How to filter data from a relation based on conditions ?
03:26
How to select one or more columns from a relation ?
07:27
How to store the result of your Pig Job into a defined data storage space ?
03:25
How to remove duplicate tuples in a relation ?
09:10
How to perform the Pig COGROUP operation ?
13:54
How to perform aggregations on input data to create data to be used by Pig ?
05:17
How to perform join of two files based on join keys ?
10:15
How to sort a relation based on one or more defined sort keys ?
02:26
How to duplicate the incoming schema into identical output flows as needed ?
02:32
How to compute the cross data of two or more relations ?
Preview
04:14
How to transform data from multiple sources to multiple targets ?
07:45
How to integrate personalized Pig Code into a Talend Job ?
05:19
Section 8: SQOOP Components
Article

Coming Soon :-)

Article

Coming Soon :-)

Article

Coming Soon :-)

Article

Coming Soon :-)

Article

Coming Soon :-)

Section 9: HCATALOG Components
04:52


08:17


04:29


03:58


02:26


Section 10: HBASE Components
05:27


04:41


00:54


07:14


05:18


Section 11: MongoDB Integration
Introduction & Installation of MongoDB
14:33
How to open a connection to a MongoDB database ?
01:59
How to close an active connection to a MongoDB database ?
01:24
How to write columns of data into a given MongoDB Collection ?
02:26

Students Who Viewed This Course Also Viewed

  • Loading
  • Loading
  • Loading

Instructor Biography

8 years experience in the area of DataStage, Talend ETL Design and Architecture, data analysis, and/or reporting, data management, modeling.

5 Years of experience in the area of Big Data.

Functional work in and familiarity with enterprise architecture concepts.

Passionate about data, technology, and innovation.

Certified Talend Open Studio for Data Integration 6.0 Consultant

Certified Talend Big Data 6.0 Developer

Certified Talend Real Time Big Data 6.0 Developer

HDP Certified Developer


Ready to start learning?
Take This Course