Video description
A step-by-step tutorial that takes you through the creation of an ETL process to populate a Kimball-style star schema
About This Video
Learn how to create ETL transformations to populate a star schema in a short span of time
Create a fully-functional ETL process using a practical approach
Follow the step-by-step instructions for creating an ETL based on a fictional company get your hands dirty and learn fast
In Detail
Companies store a lot of data, but in most cases, it is not available in a format that makes it easily accessible for analysis and reporting tools. Ralph Kimball realized this a long time ago, so he paved the way for the star schema.
Building a Data Mart with Pentaho Data Integration walks you through the creation of an ETL process to create a data mart based on a fictional company. This course will show you how to source the raw data and prepare it for the star schema step-by-step. The practical approach of this course will get you up and running quickly, and will explain the key concepts in an easy to understand manner.
Building a Data Mart with Pentaho Data Integration teaches you how to source raw data with Pentaho Kettle and transform it so that the output can be a Kimball-style star schema. After sourcing the raw data with our ETL process, you will quality check the data using an agile approach. Next, you will learn how to load slowly changing dimensions and the fact table. The star schema will reside in the column-oriented database, so you will learn about bulk-loading the data whenever possible. You will also learn how to create an OLAP schema and analyze the output of your ETL process easily.
By covering all the essential topics in a hands-down approach, you will be in the position of creating your own ETL processes within a short span of time.
Table of contents
-
Chapter 1 : Getting Started
- The Second-hand Lens Store 00:06:49
- The Derived Star Schema 00:04:30
- Setting up Our Development Environment 00:07:07
-
Chapter 2 : Agile BI – Creating ETLs to Prepare Joined Data Set
- Importing Raw Data 00:03:23
- Exporting Data Using the Standard Table Output 00:04:33
- Exporting Data Using the Dedicated Bulk Loading 00:04:32
-
Chapter 3 : Agile BI – Building OLAP Schema, Analyzing Data, and Implementing Required ETL Improvements
- Creating a Pentaho Analysis Model 00:03:26
- Analyzing Data Using Pentaho Analyzer 00:03:50
- Improving Your ETL for Better Data Quality 00:04:15
- Chpater 4 : Slowly Changing Dimensions
- Chapter 5 : Populating Data Dimension
- Chapter 6 : Creating the Fact Transformation
-
Chapter 7 : Orchestration
- Loading Dimensions in Parallel 00:06:20
- Creating Master Jobs 00:04:10
-
Chapter 8 : ID-based Change Data Capture
- Implementing Change Data Capture (CDC) 00:04:59
- Creating a CDC Job Flow 00:04:49
-
Chapter 9 : Final Touches: Logging and Scheduling
- Setting up a Dedicated DB Schema 00:01:23
- Setting up Built-in Logging 00:04:22
- Scheduling on the Command Line 00:05:30
Product information
- Title: Building a Data Mart with Pentaho Data Integration
- Author(s):
- Release date: December 2013
- Publisher(s): Packt Publishing
- ISBN: 9781782168638
You might also like
video
Python Fundamentals
51+ hours of video instruction. Overview The professional programmer’s Deitel® video guide to Python development with …
video
Getting Started with Informatica
Get up and running with Informatica - one of the most sought-after data warehousing tools in …
video
Learning Data Modeling
In this Learning Data Modeling training course, expert author Michael Blaha will teach you how to …
video
Amazon Web Services AWS LiveLessons 2nd Edition
More Than 17 Hours of Video Instruction More than 17 hours of video instruction on Amazon …