Skip to main content

DAC - EXECUTION PLAN IN DAC

Execution Plan 

An execution plan is the process of entire work carries over in DAC, which enables to organize, schedule and execute the ETL process.

An execution plan is the collection of subject areas give in DAC and its group of tasks.

Execution plan is based on subject areas and business requirements, according to the data warehouse need to be loaded. A process of data transformation includes subject areas and schedule time it depends on the type of loading the data whether full load or incremental loads.

Architecture of Execution Plans


Types of Execution Plans:
1.      Single Source Execution Plan
2.      Multi-Source Execution Plan

Single Source Execution Plan

A single-source execution plan extracts data from a single request of a single source system container, such as Oracle Ebiz or People soft or others.

Multi-Source Execution Plan

A multi – source execution plan extracts data from a single instance of a single source system container or multi source system containers as well as multi instance of a multiple source system containers.

Types of Multi – source Execution Plan

1.      Heterogeneous
 Heterogeneous execution plan which process the instance from various source systems in different time of scheduling the execution plan one after other. Even it extracts the data from single source system with different time of execution of subject areas from different place of locations.

2.    Homogeneous
 Homogeneous execution plan which process the instance from single source system in a particular times of schedule of the tasks as one after the other. It extracts the data from single source system with multiple instances at different location and time.

Extract and Load (Single)
It is the normal and simple plan of extraction and load into the target tables of data warehouse.
Having single instance of load and extract data from single source and move the data into staging, then load to Data warehouse.




Truncate process in DAC for single load and single extract condition, which truncate the tables, indexes and recreate the indexes and it will analyze the table.

Single Load and Multiple Extract

Multiple extract and single load is the process of extracting the data from single or various source systems at various time and location using the single staging and single data warehouse.
Here we have the heterogeneous and homogeneous multi source extraction, transformation and loading process will carry out.

Homogeneous:
Having one or more similar source system at different location and time loads to single end data warehouse.


 
Heterogeneous:

            Having one or more different data source system from different location and time loads data into the end data warehouse.

Truncate process in DAC for single load and multiple extract condition, which truncate the tables, indexes and recreate the indexes and it, will analyze the table. It will analysis the data source systems schedule timings and assign the priority to drop indices and analyze the tables.
Analyze Truncate Tables Only:

Purpose of “Analyze Truncate tables only”, when we running the task for a subject areas  we extract and load as “full load” as well as “Incremental Load”  it indicates the tables associated with the execution plans will be analyzed and run the tasks according to the refresh dates.

Drop/Create Indices:

Indices mean a combination of one or more indexes. It sounds plural of index.
An Index is the optional of created tables, purpose to access the data as much as quick and easy.
By creating one or more columns of the table can have ability to gain some distributed columns or set of rows of data from the original table. It is used to reduce the system heap space and disk space.Flexible in fetching the rows very speedy.

            Indexes of the tables associated with the execution plan will be dropped and re-created. When we using full load all the indexes tables will be dropped and re-created. When we using incremental load the tables where truncated will be dropped and recreated the indexes.

Analyze:

It is the task which is used to evaluate and analyze in the data warehouse after the process of truncation, drop indexes, informatica mapping and then recreation of indexes.

Full Load Always:

It indicates the ETL process to execute the data as a “full load” process.

Refresh Dates:

Refresh dates refer to the date of the last ETL process i.e., the last time data was extracted or loaded from tables in given database. DAC uses the refresh dates to determine whether to run the incremental load commands or to run full load commands and whether to truncate the target tables.

Micro ETL Execution

Micro ETL execution plans are ETL processes that you schedule at very regular intervals, such as hourly or half-hourly. They usually handle small subject areas or subsets of larger subject areas. DAC tracks refresh dates for tables in micro ETL execution plans separately from other execution plans and uses these refresh dates in the change capture process.

Comments

  1. Hi Kashif.. I have a question. Can we don specific date range load in DAC. Like from 1995 Jan1st to 2013 Dec31st ?If yes where can i set the date parameters. If no why can we set?Please let me know ASAP. Thanks

    ReplyDelete

Post a Comment

Popular posts from this blog

Contact Me

Do You have any queries ?                   If you are having any query or wishing to get any type of help related Datawarehouse, OBIEE, OBIA, OAC then please e-email on below. I will reply to your email within 24 hrs. If I didn’t reply to you within 24 Hrs., Please be patience, I must be busy in some work. kashif7222@gmail.com

Top 130 SQL Interview Questions And Answers

1. Display the dept information from department table.   Select   *   from   dept; 2. Display the details of all employees   Select * from emp; 3. Display the name and job for all employees    Select ename ,job from emp; 4. Display name and salary for all employees.   Select ename   , sal   from emp;   5. Display employee number and total salary   for each employee. Select empno, sal+comm from emp; 6. Display employee name and annual salary for all employees.   Select empno,empname,12*sal+nvl(comm,0) annualsal from emp; 7. Display the names of all employees who are working in department number 10   Select ename from emp where deptno=10; 8. Display the names of all employees working as   clerks and drawing a salary more than 3000   Select ename from emp where job=’clerk’and sal>3000; 9. Display employee number and names for employees who earn commission   Select empno,ename from emp where comm is not null and comm>0. 10

Informatica sample project

Informatica sample project - 1 CareFirst – Blue Cross Blue Shield, Maryland (April 2009 – Current) Senior ETL Developer/Lead Model Office DWH Implementation (April 2009 – Current) CareFirst Blue Cross Blue Shield is one of the leading health care insurance provided in Atlantic region of United States covering Maryland, Delaware and Washington DC. Model Office project was built to create data warehouse for multiple subject areas including Members, Claims, and Revenue etc. The project was to provide data into EDM and to third party vendor (Verisk) to develop cubes based on data provided into EDM. I was responsible for analyzing source systems data, designing and developing ETL mappings. I was also responsible for coordinating testing with analysts and users. Responsibilities: ·          Interacted with Data Modelers and Business Analysts to understand the requirements and the impact of the ETL on the business. ·          Understood the requirement and develope