Data Engineering - SSIS/ETL/Pipelines/Python/Web Scraping
What you'll learn
- Creates ETL pipelines
- Extract data from the web with Python
- Create SSIS Package
- Execute SSIS Package
- Build Web Scraping Script
- Prototype web scraping script
- Configure data source and data destination
- Clean and Transform Data
- Perform Data migration from SQL Server to Oracle
Requirements
- Basic knowledge of Python advised
- Basic knowledge of database concepts advised
Description
A data engineer is someone who creates big data ETL pipelines, and makes it possible to take huge amounts of data and translate it into insights. They are focused on the production readiness of data and things like formats, resilience, scaling, and security.
SQL Server Integration Services is a component of the Microsoft SQL Server database software that can be used to perform a broad range of data migration tasks. SSIS is a platform for data integration and workflow applications. It features a data warehousing tool used for data extraction, transformation, and loading .
ETL, which stands for extract, transform and load, is a data integration process that combines data from multiple data sources into a single, consistent data store that is loaded into a data warehouse or other target system.
An ETL pipeline is the set of processes used to move data from a source or multiple sources into a database such as a data warehouse or target databases.
SQL Server Integration Service (SSIS) provides an convenient and unified way to read data from different sources (extract), perform aggregations and transformation (transform), and then integrate data (load) for data warehousing and analytics purpose. When you need to process large amount of data (GBs or TBs), SSIS becomes the ideal approach for such workload.
Web scraping, web harvesting, or web data extraction is data scraping used for extracting data from websites. The web scraping software may directly access the World Wide Web using the Hypertext Transfer Protocol or a web browser. While web scraping can be done manually by a software user, the term typically refers to automated processes implemented using a bot or web crawler. It is a form of copying in which specific data is gathered and copied from the web, typically into a central local database or spreadsheet, for later retrieval or analysis.
Who this course is for:
- Beginners to Data Engineering
Instructor
We are experienced company that provides quality video based training .
Our courses are easy to follow and understand and will take you from
an absolute beginner with no technical skills to being efficient and confident with various technical skill like SQL and databases.
We have worked with companies of various sizes and provided consultancy services at various levels.
Thank you for learning with us and we hope your experience will be pleasant.