mastering data integration (ETL) with pentaho kettle PDI

hands on , real case studies ,tips, examples , walk trough a full project from start to end based on mySQL sakila DB.
3.7 (127 ratings) Instead of using a simple lifetime average, Udemy calculates a
course's star rating by considering a number of different factors
such as the number of ratings, the age of ratings, and the
likelihood of fraudulent ratings.
1,076 students enrolled
$40
Take This Course
  • Lectures 78
  • Contents Video: 10 hours
  • Skill Level All Levels
  • Languages English
  • Includes Lifetime access
    30 day money back guarantee!
    Available on iOS and Android
    Certificate of Completion
Wishlisted Wishlist

How taking a course works

Discover

Find online courses made by experts from around the world.

Learn

Take your courses with you and learn anywhere, anytime.

Master

Learn and practice real-world skills and achieve your goals.

About This Course

Published 11/2014 English

Course Description

    Why should i take this course
    Isn't it obvious? Don't you want to be the best ETL, pentaho kettle developer?

    General:
    The course is the outcome of my 10 year experience with IT projects and business intelligence and data integration with pentaho kettle.

    I developed the course because I want to share my knowledge with you.

    the best way to learn technological software and concepts is via an online course,
    structured by a real developer with actual experience that guide you through his (my) Path to knowledge.

    I will help you master ETL with pentaho kettle .

    What is the course about?The course is about taking you from the beginning and transfer you to a master of Pentaho kettle .

    The main dish of the course is a walk-through of a real pentaho kettle project hands on, case study, tips taking you from easy steps that becomes more and more complex, layer by layer, as you go forward. That way you can learn pentaho kettle as a beginner but also become an expert as you go along (and practice)

    Also I cover

    • the concepts of data integration
    • why we need it
    • what are the tools used today
    • data warehouse concepts

    Structure of the course
    the course is divided 4 main sections:

    Section 1: Theory and concepts of data integration in general
    (if you already an ETL developer you can skip that)

    Section 2: setting up the environment

    install and operate the data integration with pentaho kettle.
    Including database management and profiling the database as a source.
    PDI, navicat (to manage database), jdbc drivers, JRE, sakila database example, mysql and more .

    walk-through

    • pentaho kettle environment.
    • navicat (best database manager in my opinion)
    • power architect


    Section 3: the main dish

    • full data integration project with pentaho kettle
    • project overview
    • detailed design
    • step-by-step (divide and conquer)

    until the successfully end of the project. Including some 80% of the steps used by pentaho kettle in order to master data integration.
    You can see all the steps in the curriculum (it's too many to write them here)
    just for the example:

    a. Connect to various data sources (databases, files…)

    b. manipulate the data
    changing strings, dates and calculations, joins, lookups, slowly changing dimensions, consideration of when and how to use different steps.

    c. Work with variables
    d. outputs steps (bulk load , table output , update/insert , file output…)

    Section 4: wrapping up - go to production

    You will learn how to:

    1. deploy the project.
    2. make it stable by securing the solution – validation, error handling
    3. logging
    1. performance

What are the requirements?

  • basic SQL
  • database design

What am I going to get from this course?

  • develop real pentaho kettle projects
  • become master in transformation steps and jobs
  • know how to set pentaho kettle environment and deploy
  • be familiar with the most used steps of pentaho kettle
  • you will know to secure , validate , handle errors
  • check the performance and have the tools to solve issues

What is the target audience?

  • the Pentaho kettle course is meant for people who has some background with SQL syntax , Queries and database design , you dont need to be expert on that , I will guide you through
  • in case you dont know SQL at all , i suggest you take a course specific for that before you enroll to this course
  • this course is only for student who are serious in working hand on , practice and some more practice . it is not a reading or watching . you will be an expert but only if you try everything i show by your self

What you get with this course?

Not for you? No problem.
30 day money back guarantee.

Forever yours.
Lifetime access.

Learn on the go.
Desktop, iOS and Android.

Get rewarded.
Certificate of completion.

Curriculum

Section 1: Introduction
course promo intro
Preview
04:56
Section 2: Installations
05:33

The list of software we require in order to run and work with Pentaho ETL

Install mysql
02:37
01:57

JRE is required by Pentaho in order to run

03:49

This lecture shows how to install pentaho data integration

04:27

This lecture shows how to install navicat

Install sakila database (and notepad++)
08:44
01:59

this lecture shows how to install data architect - profile tool for databases

01:44

This lecture will show how to install expresso, a tool that acts as wizard for creating regular explressions

Section 3: Hands on - Pentaho
Pentaho PDI getting started
14:32
kettle variables part 1
09:35
kettle variables part 2
11:47
kettle database connection
08:34
Pentaho repositories
17:33
schema introduction
12:59
Section 4: Software Walkthroughs
Navicat walkthrough
13:49
10:11

This lecture is about Profiling database with power architect

Section 5: The Date Dimension
dim date intro
Preview
07:16
generate rows part 1
07:48
generate rows part 2
13:09
generate rows part 3
14:11
the add sequence
01:37
the select values
02:52
the mapping / string cut / string concat
09:22
the table output
10:54
the string operation
02:55
dim date summary
06:22
Section 6: dim time
dim time intro
Preview
01:17
arrange steps and create hours and minutes
08:03
the Cartesian step
Preview
05:01
Cartesian customer example
03:44
the modified java script value
08:39
the field set / filter rows / dummy steps
09:36
dim time summary
05:38
Section 7: dim staff
dim staff intro
Preview
02:08
the table input
08:55
the data grid / value mapper
08:11
consideration 1 - historical data in dimensions
03:10
consideration 2 - truncate or update table
04:23
consideration 3 - be like mike - deleted rows on dimension
Preview
05:47
Section 8: dim store
dim store intro
00:54
the database lookup
11:46
the stream lookup
09:00
the insert /update step
07:16
the system info
06:07
Section 9: dim customer
dim customer intro
01:03
control "changed data only" input
07:26
down it goes with the stream
03:35
slow changing dimension - concept
07:22
slow changing dimension - example
15:35
Section 10: dim film
dim film intro
01:25
objectives
05:32
the number range
Preview
06:10
the merge join / sort rows / value null
09:21
the denormaiser / split fields to rows
10:10
Section 11: fact rentals
fact rental intro
07:12
the inventory - film and store id
09:13
slow changing dimension on fact table
09:01
counter and date diff calculation
05:52
key date handling
08:32
the time dimension check
02:44
error handling step
04:31
Section 12: Go to production
production steps intro
03:30
the final job
03:44
kitchen batch file
05:05
schedule jobs
04:01
validation - secure the stream part 1
12:25
validation - secure the stream part 2
10:08
logging
10:41
performance
19:59
Section 13: ETL concepts and sources
what is ETL
17:46
the data warehouse concept
05:37
Analytical structure
12:33
ETL tools comparison
17:59
data sources part 1
14:11
data sources part 2
16:46
Section 14: Whats next...
need more input
08:43
this is a goodbye
03:33

Students Who Viewed This Course Also Viewed

  • Loading
  • Loading
  • Loading

Instructor Biography

Itamar Steinberg (inflow systems), MBA in the field of IT, Master of ETL

My name is Steinberg itamar,

I am in the field of information technology is for more than 15 years now,
I have a Master's degree (MBA) in information technology is an management.
My BA is also in the area of information technology from the University of Manchester.

During those years I was serving as manager and information technology expert
especially in the field of business intelligence and data integration.
all that time i used Pentaho kettle as the leading Data integration tool.

I started as a developer through team manager of development, head of the application department - responsible for all software, business processes and implementations,
also as project manager, my last role as an employee I was the CIO of the large company.

I was dealing with ERP, CRM business intelligence of course from all aspects of running a business. sales, transportation, customer service, imports, inventory, suppliers and manufacturing.

All of these areas gave me a unique perspective on business processes and how to analyze a company by looking at one large picture – that is the BI.
Of course that in order to combine all of those systems together you'll need data integration.

I have more than eight years working with pentaho kettle.
Six years ago I decided to start my own company - inflow systems and focus myself to business intelligence and data integration.
On those six years I, as CEO, was leading large business intelligence projects.
I hired several employees, very gifted, specific oriented to data integration and business intelligence. We have developed tens of projects from scratch at large companies like Alcatel Lucent (embedded solution), online gaming, binary options (stock market) and traditional businesses like food supplementary and organic food.

Today, I want to give from my knowledge to you because I believe that to share is the right way to go, during my learning phase I was reading a lot of books, struggling with the technology and I think I can make your life easier.

Ready to start learning?
Take This Course