mastering data integration (ETL) with pentaho kettle PDI
3.9 (179 ratings)
Instead of using a simple lifetime average, Udemy calculates a course's star rating by considering a number of different factors such as the number of ratings, the age of ratings, and the likelihood of fraudulent ratings.
1,384 students enrolled
Wishlisted Wishlist

Please confirm that you want to add mastering data integration (ETL) with pentaho kettle PDI to your Wishlist.

Add to Wishlist

mastering data integration (ETL) with pentaho kettle PDI

hands on , real case studies ,tips, examples , walk trough a full project from start to end based on mySQL sakila DB.
3.9 (179 ratings)
Instead of using a simple lifetime average, Udemy calculates a course's star rating by considering a number of different factors such as the number of ratings, the age of ratings, and the likelihood of fraudulent ratings.
1,384 students enrolled
Last updated 7/2016
English
Current price: $10 Original price: $120 Discount: 92% off
1 day left at this price!
30-Day Money-Back Guarantee
Includes:
  • 10 hours on-demand video
  • 1 Supplemental Resource
  • Full lifetime access
  • Access on mobile and TV
  • Certificate of Completion
What Will I Learn?
  • develop real pentaho kettle projects
  • become master in transformation steps and jobs
  • know how to set pentaho kettle environment and deploy
  • be familiar with the most used steps of pentaho kettle
  • you will know to secure , validate , handle errors
  • check the performance and have the tools to solve issues
View Curriculum
Requirements
  • basic SQL
  • database design
Description

    Why should i take this course
    Isn't it obvious? Don't you want to be the best ETL, pentaho kettle developer?

    General:
    The course is the outcome of my 10 year experience with IT projects and business intelligence and data integration with pentaho kettle.

    I developed the course because I want to share my knowledge with you.

    the best way to learn technological software and concepts is via an online course,
    structured by a real developer with actual experience that guide you through his (my) Path to knowledge.

    I will help you master ETL with pentaho kettle .

    What is the course about?The course is about taking you from the beginning and transfer you to a master of Pentaho kettle .

    The main dish of the course is a walk-through of a real pentaho kettle project hands on, case study, tips taking you from easy steps that becomes more and more complex, layer by layer, as you go forward. That way you can learn pentaho kettle as a beginner but also become an expert as you go along (and practice)

    Also I cover

    • the concepts of data integration
    • why we need it
    • what are the tools used today
    • data warehouse concepts

    Structure of the course
    the course is divided 4 main sections:

    Section 1: Theory and concepts of data integration in general
    (if you already an ETL developer you can skip that)

    Section 2: setting up the environment

    install and operate the data integration with pentaho kettle.
    Including database management and profiling the database as a source.
    PDI, navicat (to manage database), jdbc drivers, JRE, sakila database example, mysql and more .

    walk-through

    • pentaho kettle environment.
    • navicat (best database manager in my opinion)
    • power architect


    Section 3: the main dish

    • full data integration project with pentaho kettle
    • project overview
    • detailed design
    • step-by-step (divide and conquer)

    until the successfully end of the project. Including some 80% of the steps used by pentaho kettle in order to master data integration.
    You can see all the steps in the curriculum (it's too many to write them here)
    just for the example:

    a. Connect to various data sources (databases, files…)

    b. manipulate the data
    changing strings, dates and calculations, joins, lookups, slowly changing dimensions, consideration of when and how to use different steps.

    c. Work with variables
    d. outputs steps (bulk load , table output , update/insert , file output…)

    Section 4: wrapping up - go to production

    You will learn how to:

    1. deploy the project.
    2. make it stable by securing the solution – validation, error handling
    3. logging
    1. performance
Who is the target audience?
  • the Pentaho kettle course is meant for people who has some background with SQL syntax , Queries and database design , you dont need to be expert on that , I will guide you through
  • in case you dont know SQL at all , i suggest you take a course specific for that before you enroll to this course
  • this course is only for student who are serious in working hand on , practice and some more practice . it is not a reading or watching . you will be an expert but only if you try everything i show by your self
Students Who Viewed This Course Also Viewed
Curriculum For This Course
Expand All 78 Lectures Collapse All 78 Lectures 10:02:12
+
Introduction
1 Lecture 04:56
+
Installations
8 Lectures 30:50

The list of software we require in order to run and work with Pentaho ETL

What we are going to install?
05:33

Install mysql
02:37

JRE is required by Pentaho in order to run

Install JRE - java runtime
01:57

This lecture shows how to install pentaho data integration

Install pentaho data integration (kettle)
03:49

This lecture shows how to install navicat

Install navicat - mysql manager
04:27

Install sakila database (and notepad++)
08:44

this lecture shows how to install data architect - profile tool for databases

install power architect
01:59

This lecture will show how to install expresso, a tool that acts as wizard for creating regular explressions

Install expresso
01:44
+
Hands on - Pentaho
6 Lectures 01:15:00
Pentaho PDI getting started
14:32

kettle variables part 1
09:35

kettle variables part 2
11:47

kettle database connection
08:34

Pentaho repositories
17:33

schema introduction
12:59
+
Software Walkthroughs
2 Lectures 24:00
Navicat walkthrough
13:49

This lecture is about Profiling database with power architect

power architect walkthrough
10:11
+
The Date Dimension
10 Lectures 01:16:26

generate rows part 1
07:48

generate rows part 2
13:09

generate rows part 3
14:11

the add sequence
01:37

the select values
02:52

the mapping / string cut / string concat
09:22

the table output
10:54

the string operation
02:55

dim date summary
06:22
+
dim time
7 Lectures 41:58

arrange steps and create hours and minutes
08:03


Cartesian customer example
03:44

the modified java script value
08:39

the field set / filter rows / dummy steps
09:36

dim time summary
05:38
+
dim staff
6 Lectures 32:34

the table input
08:55

the data grid / value mapper
08:11

consideration 1 - historical data in dimensions
03:10

consideration 2 - truncate or update table
04:23

+
dim store
5 Lectures 35:03
dim store intro
00:54

the database lookup
11:46

the stream lookup
09:00

the insert /update step
07:16

the system info
06:07
+
dim customer
5 Lectures 35:01
dim customer intro
01:03

control "changed data only" input
07:26

down it goes with the stream
03:35

slow changing dimension - concept
07:22

slow changing dimension - example
15:35
+
dim film
5 Lectures 32:38
dim film intro
01:25

objectives
05:32


the merge join / sort rows / value null
09:21

the denormaiser / split fields to rows
10:10
4 More Sections
About the Instructor
Itamar Steinberg (inflow systems)
3.9 Average rating
181 Reviews
1,662 Students
2 Courses
MBA in the field of IT, Master of ETL

My name is Steinberg itamar,

I am in the field of information technology is for more than 15 years now,
I have a Master's degree (MBA) in information technology is an management.
My BA is also in the area of information technology from the University of Manchester.

During those years I was serving as manager and information technology expert
especially in the field of business intelligence and data integration.
all that time i used Pentaho kettle as the leading Data integration tool.

I started as a developer through team manager of development, head of the application department - responsible for all software, business processes and implementations,
also as project manager, my last role as an employee I was the CIO of the large company.

I was dealing with ERP, CRM business intelligence of course from all aspects of running a business. sales, transportation, customer service, imports, inventory, suppliers and manufacturing.

All of these areas gave me a unique perspective on business processes and how to analyze a company by looking at one large picture – that is the BI.
Of course that in order to combine all of those systems together you'll need data integration.

I have more than eight years working with pentaho kettle.
Six years ago I decided to start my own company - inflow systems and focus myself to business intelligence and data integration.
On those six years I, as CEO, was leading large business intelligence projects.
I hired several employees, very gifted, specific oriented to data integration and business intelligence. We have developed tens of projects from scratch at large companies like Alcatel Lucent (embedded solution), online gaming, binary options (stock market) and traditional businesses like food supplementary and organic food.

Today, I want to give from my knowledge to you because I believe that to share is the right way to go, during my learning phase I was reading a lot of books, struggling with the technology and I think I can make your life easier.