Professional Documents
Culture Documents
TRANSFORMATIONS:
AGGREGATOR:(ACTIVE AND CONNECTED)-IT IS USED TO PERFORM
AGGREGATE CALCULATIONS.
PROPERTIES:
1)CACHE DIRECTORY:-LOCAL DIRECTORY WHICH CRAETES INDEX AND
DATA CACHE.
2)TRACING LEVEL:-AMOUNT OF DETAIL DISPLAYED IN SESSION LOG
FILE.
3)SORTED INPUT:-INDIACATES INPUT DATA IS PRE-SORTED BY
GROUPS.SELECT THIS OPTION ONLY IF THE MAPPING PASSES SORTED
DATA TO THE AGGREGATOR TRANSFORMATION.
4)DATA CHACHE DEFAULT SIZE IS 2MB.
5)INDEX CACHE DEFAULT SIZE IS 1MB.
-->IF SOURCE DATA IS GREATER THAN SIZE OF CHACHE THEN
INTEGRATION SERVICE WILL USE DISK SPACE.
6)TRANSFORMATION SCOPE:SPECIFIES HOW THE INTEGRATION SERVICE
APPLIES THE TRANSFORMATION LOGIC TO INCOMING DATA.
7)TRANSACTION:APPLIES THE TRANSFORMATION LOGIC TO ALL ROW IN
A TRANSACTION.
COMPONENTS&OPTIONS:
->AGGREGATE CACHE,AGGREGATE EXPRESSION AND GROUP BY PORT.
================================================
================================
EXPRESSION TRANSFORMATION:(PASSIVE&CONNECTED)
USING THE EXPRESSION TRANSFORMATION WE CAN CALCULATE VALUES
IN A SINGLE ROW.
FOR EXAMPLE TO CANCAT THE FIRST NAME,LAST NAME AND WE CAN
ALSO USE TO TEST CONDIONAL STATEMENTS AND ALSO WE CAN DO
ROW LEVEL CALCULATIONS.
================================================
================================
FILTER TRANSFORMATION:(ACTIVE&CONNECTED)
Filter
================================================
==================================
PROPERTIES:
1)CASE SENSITIVE STRING CAMPARISION:IF SELECTED THE INTEGRATION
SERVICE USES CASE SENSITIVE STRING COMPARISONS WHEN
PERFORMING JOINS ON STRING COLUMNS.
2)CACHE DIRECTORY:SPECIFIES THE DIRECTORY USED TO CACHE
MASTER OR DETAIL ROWS AND THE INDEX TO THESE ROWS.
3)JOIN TYPE:SPECIFIES THE TYPE OF JOIN-->NORMAL,MASTER
OUTER,DETAIL OUTER,FULL OUTER.
4)NULL ORDERING IN MASTER:DETERMINE WETHER NULL IS THE
HIGHEST OR LOWEST VALUE IN THE MASTER RELATION.
5)NULL ORDERING IN DETAIL:DETERMINE WETHERNULL IS THE HIGHEST
OR LOWEST VALUE IN THE DETAIL RELATION.
6)SORTED INPUT:SPECIFIES THE DATA IS SORTED,USING SORTED INPUT
WILL IMPROVE PERFORMANCE.
================================================
==================================
LOOKUP(PASSIVE&CONNECTED,UN-CONNECTED)
THIS TRANSFORMATION IS USED TO LOOKUP DATA IN RELATED TABLE
OR FLATFILE OR VIEW OR SYNONYM.
->WE CAN USE MULTIPLE LOOK UP IN ONE MAPPING.
Unconnected Lookup
2)CONNECTION INFO
3)CALL TEXT:-STORED PROCEDURE NAME
4)STORED PROCEDURE TYPE---->TARGET PRE-LOAD,TARGET POSTLOAD,NORMAL,SOURCE PRE-LOAD,SOURCE POST LOAD.
5)EXECUTION ORDER:-IF WE HAVE MORE THAN ONE STORED
PROCEDURE THEN WE CAN SPECIFY ORDER.
--->WE CAN USE UNCONNECTED WHERE WE WANT USE PRE OR POST ON
SOURCE AND TARGET.
--->WE CAN USE UN-CONNECTED WHERE WE WANT SOME INPUT AND
RETURN SOME OUTPUT THERE WE USE THIS.
================================================
==================================
TRANSACTION CONTROL(ACTIVE AND CONNECTED)
->POWER CENTRE LETS YOU CONTROL COMMIT AND ROLL BACK
TRANSACTIONS BASED ON SET OF ROWS THAT PASSES THROUGH A
TRANSACTIONAL CONTROL TRANSFORMATION.
->WE CAN DEFINE TRANSACTION BASED ON VARYING NUMBER OF
INPUTS ROWS.
->WE CAN DEFINE TRANSACTIONAL CONTROL IN MAPPING LEVEL AND
SESSION LEVEL.
->WHEN YOU RUN THE SESSION,THE INTEGRATION SERVICE EVALUTES
THE EXPRESSIONS FOR EACH ROW THAT ENTERS THE
TRNSACTIONS.WHEN IT EVALUTES A COMMIT ROW,IT COMMITS ALL
ROWS IN THE TRANSACTION TO THE TARGETS.WHEN INTEGRATION THE
INTEGRATION SERVICE EVALUTES ROLL BACK ROW,IT ROLL BACKS IN
THE TRANSACTION FROM THE TARGET
->FOLLOWING BUILT IN VARIABLES,WE WILL USE IN EXPRESSION
EDITOR.WHEN YOU CREATED TRANSACTIONAL CONTROL
-->TC_CONTINUE_TRANSACTION:-INTEGARATION SERVICE DOES NOT
PERFORM ANY CHANGE FOR THIS ROW
-->TC_COMMIT_BEFORE:-IT COMMITS,BEFORE NEW TRANSACTION COME
AND WRITES CURRRENT ROW TO THE TARGET.
-->TC_COMMIT_AFTER:-->TC_ROLLBACK_BEFORE:-->TC_ROLLBACK_AFTER:-
2) There was one more scenario where I got very good performance in
the mapping level .Rather than using lookup transformation if we can
able to do outer join in the source qualifier query override this will give
you good performance if both lookup table and source were in the same
database. If lookup tables is huge volumes then creating cache is costly.
3) And also if we can able to optimize mapping using less no of
transformations always gives you good performance.
4) If any mapping taking long time to execute then first we need to look in
to source and target statistics in the monitor for the throughput and also
find out where exactly the bottle neck by looking busy percentage in the
session log will come to know which transformation taking more time ,if
your source query is the bottle neck then it will show in the end of the
session log as query issued to database that means there is a
performance issue in the source query.we need to tune the query using
Now Lookup can be configured as an active transformation - it can return multiple rows on
successful match
Now you can write SQL override on un-cached lookup also. Previously you could do it only
on cached lookup
You can control the size of your session log. In a real-time environment you can control the
session log file size or time
Database deadlock resilience feature - this will ensure that your session does not
immediately fail if it encounters any database deadlock, it will now retry the operation again. You
can configure number of retry attempts.
4-->Client Tools access the repository via that centralized machine, resources are distributed
dynamically.
5-->It has a support for unstructured data which includes spreadsheets, email, Microsoft Word files,
presentations and .PDF documents. It provides high availability, seamless fail over, eliminating single
points of failure.
6-->Informatica has now added more tightly integrated data profiling, cleansing, and matching
capabilities.
7-->A new web based administrative console has been added.
8-->Java transformation is introduced.
9-->Ability to write a Custom Transformation in C++ or Java.
10-->Dynamic configuration of caches and partitioning
11-->New transforms like SQL transform, HTTP transform, Salesforce, etc. have been added in Version
8.
DTM Process
The Load Manager creates one DTM process for each
session
in the workflow. It performs the following tasks:
Reads session information from the repository.
Expands the server, session, and mapping variables and
parameters.
Creates the session log file.
Validates source and target code pages.
2. Terse: Integration Service logs initialization information and error messages and
notification of rejected data.