You are on page 1of 19

INFORMATICA NOTES:

TRANSFORMATIONS:
AGGREGATOR:(ACTIVE AND CONNECTED)-IT IS USED TO PERFORM
AGGREGATE CALCULATIONS.
PROPERTIES:
1)CACHE DIRECTORY:-LOCAL DIRECTORY WHICH CRAETES INDEX AND
DATA CACHE.
2)TRACING LEVEL:-AMOUNT OF DETAIL DISPLAYED IN SESSION LOG
FILE.
3)SORTED INPUT:-INDIACATES INPUT DATA IS PRE-SORTED BY
GROUPS.SELECT THIS OPTION ONLY IF THE MAPPING PASSES SORTED
DATA TO THE AGGREGATOR TRANSFORMATION.
4)DATA CHACHE DEFAULT SIZE IS 2MB.
5)INDEX CACHE DEFAULT SIZE IS 1MB.
-->IF SOURCE DATA IS GREATER THAN SIZE OF CHACHE THEN
INTEGRATION SERVICE WILL USE DISK SPACE.
6)TRANSFORMATION SCOPE:SPECIFIES HOW THE INTEGRATION SERVICE
APPLIES THE TRANSFORMATION LOGIC TO INCOMING DATA.
7)TRANSACTION:APPLIES THE TRANSFORMATION LOGIC TO ALL ROW IN
A TRANSACTION.
COMPONENTS&OPTIONS:
->AGGREGATE CACHE,AGGREGATE EXPRESSION AND GROUP BY PORT.
================================================
================================
EXPRESSION TRANSFORMATION:(PASSIVE&CONNECTED)
USING THE EXPRESSION TRANSFORMATION WE CAN CALCULATE VALUES
IN A SINGLE ROW.
FOR EXAMPLE TO CANCAT THE FIRST NAME,LAST NAME AND WE CAN
ALSO USE TO TEST CONDIONAL STATEMENTS AND ALSO WE CAN DO
ROW LEVEL CALCULATIONS.
================================================
================================
FILTER TRANSFORMATION:(ACTIVE&CONNECTED)

IT IS ACTIVE SO,THE FILTER TRANSFORMATION MAY CHANGE THE


NUMBER OF ROWS PASSED THROUGH IT.
=>IT PASS THE RECORDS WHICH MEET THE CONDITION.
=>IF ANY ROW NOT MEET THAT CONDITION THEN IT WILL DISCARD
THAT RECORD.
=>KEEP THE FILTER TRANSFORMATION AS CLOSE AS SOURCE QUALIFIER
BECAUSE IF WE FILTER RECORDS WHICH ARE UNWANTED THEN
PERFORMANCE WILL BE IMPROVED.
PROPERTIES OF FILTER:
->FILTER CONDITION IS AN EXPRESSION THET RETURNS TRUE OR
FALSE'ENTER CONDITIONS USING THE EXPRESSION EDITOR AVAILABLE
ON PROPERTIES TAB.
================================================
==================================
ROUTER TRANSFORMATION:(ACTIVE AND CONNECTED)
->ROUTER TRANSFORMATION USES CONDITION TO TEST DATA
->IT TEST ONE OR MORE CONDITIONS
->IT ROUTES THE DATA TO DEFAULT GROUP IF DATA DOESNOT MEET ANY
CONDTION.
->HERE INTEGRATION SERVICE PROCESS INCOMING DATA FOR ONCE.
->ROUTER HAVING INPUT GROUP AND OUTPUT GROUP
WE HAVE TO CREATE NUMBER OF GROUPS WITH CONDITION THERE WILL
BE A DEFAULT GROUP GROUP FOR EVERY ROUTER TRANSFORMATION.

What is the difference between Router and Filter?


Router

Filter

Router transformation divides the incoming


records into multiple groups based on some
condition. Such groups can be mutually
inclusive (Different groups may contain
same record)

Filter transformation restricts or blocks the


incoming record set based on one given
condition.

Router transformation itself does not block


any record. If a certain record does not
match any of the routing conditions, the

Filter transformation does not have a default


group. If one record does not match filter

record is routed to default group

condition, the record is blocked

Router acts like CASE.. WHEN statement in


SQL (Or Switch().. Case statement in C)

Filter acts like WHERE condition is SQL.

================================================
==================================

JOINER TRANSFORMATION:(ACTIVE AND CONNECTED)


USING JOINER TRANSFORMATION WE CAN JOIN TWO HETEROGENEOUS
SOURCES RESIDING IN DIFFERENT LOCATIONS OR FIE SYSTEMS.
->ATLEAST ONE MATCHING ROW SHOULD BE COMMON IN TWO TABLES.U
HAVE TO SPECIFY THAT MATCHING COLOUMN IN CONDITION.
-->WHY JOINER IS A BLOCKING TRANSFORMATION?
BECAUSE TO ENSURE THE MASTER RECORDS IT WILL BLOCK THE DATAIL
TABLE RECORDS.

PROPERTIES:
1)CASE SENSITIVE STRING CAMPARISION:IF SELECTED THE INTEGRATION
SERVICE USES CASE SENSITIVE STRING COMPARISONS WHEN
PERFORMING JOINS ON STRING COLUMNS.
2)CACHE DIRECTORY:SPECIFIES THE DIRECTORY USED TO CACHE
MASTER OR DETAIL ROWS AND THE INDEX TO THESE ROWS.
3)JOIN TYPE:SPECIFIES THE TYPE OF JOIN-->NORMAL,MASTER
OUTER,DETAIL OUTER,FULL OUTER.
4)NULL ORDERING IN MASTER:DETERMINE WETHER NULL IS THE
HIGHEST OR LOWEST VALUE IN THE MASTER RELATION.
5)NULL ORDERING IN DETAIL:DETERMINE WETHERNULL IS THE HIGHEST
OR LOWEST VALUE IN THE DETAIL RELATION.
6)SORTED INPUT:SPECIFIES THE DATA IS SORTED,USING SORTED INPUT
WILL IMPROVE PERFORMANCE.
================================================
==================================

LOOKUP(PASSIVE&CONNECTED,UN-CONNECTED)
THIS TRANSFORMATION IS USED TO LOOKUP DATA IN RELATED TABLE
OR FLATFILE OR VIEW OR SYNONYM.
->WE CAN USE MULTIPLE LOOK UP IN ONE MAPPING.

TASK'S PERFORMED BY LOOKUP:


1)TO GET RELATED VALUE
2)TO PERFORM A CALCULATION
3)UPDATE SCD

LOOKUP PERFORM FOLLOWING TYPES OF LOOKUPS:


1)RELATIONAL OR FLATFILE LOOKUP:PERFORM LOOKUP ON A FLAT FILE
OR RELATIONAL TABLE.
2)PIPELINE LOOKUP:PERFORM A LOOKUP ON APPLICATION SOURCES
LIKE SAP
3)CONNECTED LOOKUP:A CONNECTED LOOKUP TRANSFORMATION
RECIEVES SOURCE DATA,PERFORM LOOKUP AND RETURN A DATA TO A
PIPELINE.IT CAN RETURN MULTIPLE COLUMNS.
4)UNCONNECTED LOOKUP IS NOT CONNECTED TO A SOURCE OR
TARGET,IT WILL GET INPUT FROM RESULT OF :LKP EXPRESSION.

What are the differences between Connected and Unconnected Lookup?


Connected Lookup

Unconnected Lookup

Connected lookup participates in dataflow


and receives input directly from the pipeline

Unconnected lookup receives input values


from the result of a LKP: expression in
another transformation

Connected lookup can use both dynamic and


static cache

Unconnected Lookup cache can NOT be


dynamic

Connected lookup can return more than one


column value ( output port )

Unconnected Lookup can return only one


column value i.e. output port

Connected lookup caches all lookup columns

Unconnected lookup caches only the lookup

output ports in the lookup conditions and the


return port

Supports user-defined default values (i.e.


value to return when lookup conditions are
not satisfied)

Does not support user defined default values

What r the types of lookup caches?


1)STATIC CACHE:IT WILL NOT CHANGE DURING THE SESSION, I MEAN IT
REMAINS SAME.
BY DEFAULT LOOKUP CACHE IS STATIC.
2)DYNAMIC CACHE:INTEGRATION SERVICE DYNAMICALLY INSERTS OR
UPDATES DATA IN THE LOOK UP CACHE.
3)PERSISTANT CACHE:WE CAN SAVE LOOKUP CACHE FILES AND REUSE
THEM WHEN NEXT TIME INTEGRATION SERVICE PROCESSES LOOKUP
TRANSFORMATION,CONFIGURED TO USE CACHE.
4)SHARED CACHE:YOU CAN SHARE LOOKUP CACHE BETWEEN MULTIPLE
LOOKUP TRANSFORMATION,AS NAMED CAN BE USED IN MULTIPLE
MAPPING WHERE AS UN NAMED CAN USE WITH IN SAME MAPPING.
PROPERTIES OF LOOKUP:
1)LOOKUP SQL OVERRIDE:WE CAN OVERRIDE THE DEFAULT SQL
STATEMENT TO QUERY UP THE LOOKUP TABLE.
2)LOOKUP TABLE NAME:SPECIFY THE TABLE NAME AND CONNECTION
INFORMATION.
3)LOOKUP SOURCE FILTER:RESTRICTS THE LOOKUPS
4)LOOKUP CACHE IS ENABLED
5)LOOK UP POLOCY ON MULTIPLE MATCH: WHEN IT FINDSMULTIPLE
MATCH, WE HAVE TO MENTION AS
->USE FIRST VALUE
->USE LAST VALUE
->USE ANY VALUE
->REPORT ERROR
6)LOOKUP CONDITION:DISPLAYS THE LOOKUP CONDITION WHICH WE
SET IN THE CONDITION TAB.
7)CONNECTION INFORMATION

8)SOURCE TYPE:WE HAVE TO MATION WHAGT TYPE OF SOURCE WETHER


IT IS RELATIONAL OR FLAT FILE.
9)LOOKUP CACHE DIRECTORY NAME:$PMCACHEDIR--DIRECTORY NAE
WHERE LOOKUP SOURCE RESIDING.
10)LOOKUP CACHE PERSISTANT:IF YOU ENABLE THIS THAN,AFTER
SESSION COMPLETED THEN INTEGRATION SERVICE WILL SAVE THAT
CACHE FILE.
11)DYNAMIC LOOKUP CACHE:IF YOU SELECT,INSERT OR UPDATES ROWS
IN THE LOOKUP CACHE AS IT PASSES ROWS TO THE TARGET TABLE.
12)INSERT ELSE UPDATE:WHEN THIS IS ENABLE,THE INTEGRATION
SERVICE INSERTS NEW ROWS IN THE CACHE AND UPDATE EXISTING
ROWS IN THE CACHE.
13)UPDATE ELSE INSERT:IT WILL UPDATE IF ROW EXISTED,IF NOT EXISTS
THEN INSERT.
================================================
==================================

NORMALIZER TRANSFORMATION:(ACTIVE AND CONNECTED)


->USING NORMALIZER WE CAN CONVERT NUMBER OF COLUMNS INTO
ROWS.
->IT GENERATES KEY SEQUENCE NUMBER EACH TIME IT PROCESSES A
SINGLE ROW.
->WE CREATE VSAM NORMALIZER WHICH IS NON-REUSABLE AND IT ACT
LIKE SOURCE QUALIFIER FOR THE COBOL SOURCE.
->VSAM IS NOT REUSABLE
->RESET GENERATED KEY SEQUENCE VALUES AT THE END OF SESSION.
->START THE GENERATED KEY SEQUENCE VALUES FROM 1
->OCCURS:NUMBER OF INSTANCE OF COLUMNS.
================================================
==================================
RANK TRANSFORMATION:(ACTIVE AND CONNECTED)
->USING RANK TRANSFORMATION WE CAN SELECT TOP OR BOTTUM
RANK OF DATA.

->WE CAN ALSO RETURNS THE STRINGS AT THE TOP OT BOTTUM OF A


SESSION SORT ORDER.
->INTEGRATION SERVICE CACHES INPUT DATA UNTILL IT CAN PERFORM
RANK CALCULATOR.
================================================
==================================
SEQUENCE GENERATOR:(PASSIVE AND CONNECTED)
IT GENERATES NUMERIC VALUES.WE CAN CREATE PRIMARY KEY VALUES
(OR) CYCLE THROUGH A SEQ RANGE OF NUMBERS
->IF YOU CONNECT NEXT VAL AS INPUT TO ANOTHER TRANSFORMATION
THEN IT WILL GIVE SEQ OF NUMBERS.
->IF YOU CONNECT CURR VAL TO NEXT TRANSFORMATION ,THEN IT WILL
GENERATE SAME NUMBER TO ALL RECORDS.
PRPERTIES:
1)START VALUE:BY DEFAULT IT IS 0 OR WE CAN CHANGE.
2)INCREMENT BY:DIFFERENCE BETWEEN TWO CONSECUTIVE VALUES
FROM THE NEXT VAL PORT
3)END VALUE:MAXIMUM VALUE THE INTEGRATION SERIVICE
GENERATES,IF WE NOT COFIGURE TO CYCLE THEN I MAX VALUE IS
REACHED THEN SESSION FAILS.
4)CURRENT VALUE
5)CYCLE:IF ENABLE ,THEN INTEGRATION SERVICE CYCLE THROUGH THE
SEQ RANGE.
SORTER TRANSFORMATION:(ACTIVE AND CONNECTED)
->WE CAN SORT DATA USING SORTER TRANSFORMATION.
->WE CAN SORT IN ASC OR DESC ACCORDING TO SPECIFIED KEY.
->WE CAN ALSO CONFIGURE SORTER TRANSFORMATION FOR CASE
SENSITIVE.
->WE CAN ALSO SPECIFY WETHER THE OUTPUT ROWS SHOULD BE
DISTINCT.
================================================
=================================
SOURCE QUALIFIER(ACTIVE AND CONNECTED)

THE SOURCE QUALIFIER REPRESENTS THE ROWS THAT THE


INTEGRATION SERVICE READS WHEN IT RUN A SESSION.
->SOURCE QUALIFIR COME BY DEFAULT WITH RELATIONAL AND FLAT
FILES.
PROPERTIES:
1)SQL-QUERY:WE CAN OVERRIDE THE DEFAULT SQL QUERY WHICH IS
GENERATED IN SQ
2)USER DEFINED JOIN:WE CAN JOIN DATA FROM MULTIPLESOURCES.
3)SOURCE FILTER:IF WE SPECIFY THE FILTER CONDITION THEN THE
INTEGRATION SERVICE APLLIES WHILE QURIYING ROWS.
4)NO. OF SORTED PORTS:WE CAN INDICATE THE NO OF PORTS AS
SORTED BASED ON THET THAT INTEGRATION SERVICE ADDS AN ORDER
BY CLAUSE TO DEFAULT QUERY.
5)SELECT DISTINCT:IF WE SELECT THIS THEN IT WILL ADD DISTINCT
STATEMENT TO DEFAULT QUERY.
6)PRE-SQL:SQL COMMANDS TO RUN AGAINST THE SOURCE DATABASE
BEFORE INTEGRATION SERVICE READS THE SOURCE.
7)POST SQL: SQL COMMANDS TO RUN AGAINST THE SOURCE DATABASE
AFTER INTEGRATION SERVICE READS THE SOURCE
================================================
==================================
STORED PROCEDURE(PASSIVE &CONNECTED/UN-UNCONNECTED)
->STORED PROCEDURE IS AN IMPORTANT TOOL FOR POPULATING AND
MANTAINING DATABASES.
->DATABASE ADMIN CREATES STORED PROCEDURE TO AUTOMATE
TASTES THAT ARE TOO COMPLICATED FOR STANDARD SQL STATEMENTS.
TASKS CAN PERFORM BY STORED PROCEDURE
->CHECK THE STATUS OF THE TARGET DATABASE BEFORE LOADING INTO
IT.
->TO CHECK WETHER ENOUGH SPACE EXISTS IN A DATABASE
->PERFORM A SPECIFIED CALCULATION.
->DROP AND RE-CREATE INDEXES.
PROPERTIES:1)STORED PROCEDURE NAME

2)CONNECTION INFO
3)CALL TEXT:-STORED PROCEDURE NAME
4)STORED PROCEDURE TYPE---->TARGET PRE-LOAD,TARGET POSTLOAD,NORMAL,SOURCE PRE-LOAD,SOURCE POST LOAD.
5)EXECUTION ORDER:-IF WE HAVE MORE THAN ONE STORED
PROCEDURE THEN WE CAN SPECIFY ORDER.
--->WE CAN USE UNCONNECTED WHERE WE WANT USE PRE OR POST ON
SOURCE AND TARGET.
--->WE CAN USE UN-CONNECTED WHERE WE WANT SOME INPUT AND
RETURN SOME OUTPUT THERE WE USE THIS.
================================================
==================================
TRANSACTION CONTROL(ACTIVE AND CONNECTED)
->POWER CENTRE LETS YOU CONTROL COMMIT AND ROLL BACK
TRANSACTIONS BASED ON SET OF ROWS THAT PASSES THROUGH A
TRANSACTIONAL CONTROL TRANSFORMATION.
->WE CAN DEFINE TRANSACTION BASED ON VARYING NUMBER OF
INPUTS ROWS.
->WE CAN DEFINE TRANSACTIONAL CONTROL IN MAPPING LEVEL AND
SESSION LEVEL.
->WHEN YOU RUN THE SESSION,THE INTEGRATION SERVICE EVALUTES
THE EXPRESSIONS FOR EACH ROW THAT ENTERS THE
TRNSACTIONS.WHEN IT EVALUTES A COMMIT ROW,IT COMMITS ALL
ROWS IN THE TRANSACTION TO THE TARGETS.WHEN INTEGRATION THE
INTEGRATION SERVICE EVALUTES ROLL BACK ROW,IT ROLL BACKS IN
THE TRANSACTION FROM THE TARGET
->FOLLOWING BUILT IN VARIABLES,WE WILL USE IN EXPRESSION
EDITOR.WHEN YOU CREATED TRANSACTIONAL CONTROL
-->TC_CONTINUE_TRANSACTION:-INTEGARATION SERVICE DOES NOT
PERFORM ANY CHANGE FOR THIS ROW
-->TC_COMMIT_BEFORE:-IT COMMITS,BEFORE NEW TRANSACTION COME
AND WRITES CURRRENT ROW TO THE TARGET.
-->TC_COMMIT_AFTER:-->TC_ROLLBACK_BEFORE:-->TC_ROLLBACK_AFTER:-

PROPERTY:-TRS ANSACTIONAL CONTROL CONDITION.


================================================
==================================
UNION TRANSFORMATION(ACTIVE AND CONNECTED)
->UNION TRANSFORMATION IS USED TO MERGE DATA FROM MULTIPLE
SOURCES.
->PROCESS ALL INPUT GROUPS PARALALLY.
->IT READS CONCURRANTLY.
->IT PROCESS THE BLOCKS OF DATA BASED ON THE ORDER IT RECEIVES.
->WE CAN MERGE DATA FROM TABLE,WHOSE HAVING SAME STRUCTURE
OTHER WISE WE USE JOINER.
================================================
==================================
UPDATE STRATEGY(CONNECTED AND ACTIVE)
->WE WILL USE UPDATE STRATEGY WHEN WE ARE APPLYNG SCD'S OR
TYPES.
WE CAN DO THIS ONE IN SESSION LEVEL AND MAPPING LEVEL.
->WITH0202 IN SESSION:-USE TREAT ROWS AS INSERT,UPDATE,DELETE
AND DATA DRIVEN
->WITH IN MAPPING:-UPDATE STRATEGY TRANSFORMATION TO FLAG
ROW FOR INSERT,DELETE,UPDATE OR REJECT.
PROPERTIES:1)UPDATE STRATEGY EXPRESSION:DD_INSERT,DD_UPDATE,DD_DELETE,DD_REJECT
0---INSERT
1---UPDATE
2--DELETE
3---REJECT.
================================================
==================================

HOW CAN A TRANSFORMATION BE MADE REUSABLE?


IN THE EDIT PROPERTIES OF ANY TRANSFORMATION THERE IS A CHECK
BOX TO MAKE IT REUSABLE,BY CHECKING THAT IT BECOMES
REUSABLE.YOU CAN EVEN CREATE REUSABLE TRANSFORMATION IN
TRANSFORMATION DEVELOPER.
WHAT IS MAPPLET AND HOW IS IT DIFFERENT FROM REUASABLE
TRANSFORMATION?
A MAPPLET IS A REUSABLE OBJECT THAT REPRESENTS THE SET OF
TRANSFORMATIONS.IT ALLOWS YOU TO REUSE TRANSFORMATION LOGIC
AND CAN CONTAIN AS MANY TRANSFORMATION AS YOU NEED.
IN MAPPLET WE CAN CREATE SET OF TRANSFORMATIONS WHERE AS
REUSABLE TRANSFORMATION IS SINGLE.
WHAT IS A SESSION?
A SESSION IS A SET OF INSTRUCTIONS THAT TELL THE INFORMATICA
SERVER HOW AND WHEN TO MOVE DATA FFROM SOURCES TO TARGET.
WHAT ARE THE DML COMMLANDS POSSIBLE THROOUGH AN UPDATE
STRATEGY?
DD_INSERT-0,DD_UPDATE-1,DD_DELETE-2,DD_REJECT-3
INFORMATICA PROVIDES 10 TYPES OF TASKS:1)SESSION:- A SESSION IS A SET OF INSTRUCTIONS THAT TELL THE
INFORMATICA SERVER HOW AND WHEN TO MOVE DATA FFROM SOURCES
TO TARGET.
2)COMMLAND:-THI S TASK ALLOWS US TO SPECIFY ONE OR MORE SHELL
COMMANDS IN UNIX OR DOS COMMANDS IN WINDOWS TO RUN DURING
THE WORKFLOW.
3)EMAIL TASK:-THIS PROVIDES US TO SENDEMAIL DURING THE
WORKFLOW.
4)DECISION:-THIS IS USED TO CHECK CONDITION AND BASED ON THE
OUTCOME OF THE CONDITIONS,WE CAN CONTROL FLOW OF THE
PROCESS.
5)CONTROL:-THIS TASK IS USED TO DECIDE WHEN THE PROCESS
SHOULD BE STOPPED.
THERE ARE THREE TYPES TO CONTROL A PROCESS
1)FAIL:-PROCESS IS NOT STOPPED BUT AFTER THE COMPLETION IT IS
MARKED AS 'FAILED'.

2)STOP:-DATA QURY IS STOPPED IMMEDIATELY BUT PROCESS


CONTINUES FOR THE DATA WHICH IS ALREADY LOADED INTO BUFFER.
3)ABORT:-SAME AS STOP BUT IN THIS CASE ONLY 60 SEC'S IS ALLOWED
FOR COMPLETION OF PROCESS FOR BUFFER DATA.
6)TIMER:-IT IS USED TO DELAY THE PROCESS FOR SPECIFIC TIME
PERIOD.
7)ASSIGNMENT:-IT IS USED TO ASSIGN VALUES INTO VARIABLES
DEFINED AT WORKFLOW LEVEL
8)EVENT-WAIT:-IT IS USED TO DEFINE USER-DEFINED OR PRE-DEFINED
EVENTS,THE EVENTS ARE USED TO DELAY THE PROCESS.
9)EVENT-RAISE:-IT IS USED TO FIRE A USER-DEFINED EVENT.
10)WORKLET:-IT IS SAME AS WORK FLOW BUT RE-USABLE.
WHEN YOU ARE LOADING DATA WHAT TYPE OF ERRORS HAVE YOU
FACED?
->NULL VALUES
->CALCULATION ARE IN-CORRECT
->STOP ON ERRORS
->COMMIT INTERVAL
WHAT ARE 2 MODES USED FOR DATA MOVEMENT IN INFORMATICA?
1)UNI-CODE:-IT ALLOWS 2 BYTES TO EACH CHAR AND USES ADDITIONAL
BYTE FOR EACH
NON-ASCII CHARACTER
2)ASCII:-IT HOLDS ALL DATA IN A SINGLE BYTE.
HOW TO LOAD FIRST AND LAST RECORD INTO THE TARGET?
FIRST RECORD CAN BE LOADED INTO TARGET USING RANK
TRANSFORMATIOM(TOP)
LAST RECORD CAN BE DONE USING AGGREGATOR WITHOUT GROUP BY
OPTION.
WHAT ARE THE SITUATIONSTHEN SESSION CAN FAIL?
->SERVER CANNOT ALLOCATE ENOUGH RESOURCES
->WHEN LOCK EXECUTES ON TABLE
->UNABLE TO EXECUTE POST AND PRE LOAD STORED PROCEDURES.
->DATABASE ERRORS

PERFORMANCE TUNING IN INFORMATICA:1)THE PERFORMANCE OF INFORMATICA SERVER IS RELATED TO


NETWORK CONNECTION,
DATA MOVES ACROSS AT LESS THAN 1MB/SEC.WHERE AS LOCAL DISK
MOVES DATA 5-20 TIMES FASTER.THUS NETWORK CONNECTION EFFECT
PERFORMANCE,SO AVOID NETWORK CONNECTIONS.
2)FLAT-FILE:-IF IT STORES IN OTHER MACHINE->MOVE TO LOCAL
MACHINE THAT CONSISTS OF INFORMATICA SERVER..WHICH IMPROVES
THE SPEED.
3)JOIN MULTIPLE SOURCE TABLE IN SOURCE QUAALIFIER->SAME DATA
BASE,SQL OVERRIDE LIKE GROPUP BY AND ORDER BY..
4)BY ENABLING LOOKUP CACHE IF SESSION CONTAINS LOOK UP
TRANSFORMATIONS.
5)USE SORTED INPUT WHEN YOU ARE USING AGGREGATOR,RANK AND
JOINER TRANSFORMATION.
WHAT IS LOOKUP OVERIDE?
THIS FEATURE IS SIMILAR TO ENTERING A CUSTOM QUERY IN A SOURCE
QUALIFIER TRANSFORMATION.WHEN ENTERING A LOKKUP SQL
OVERIDE,YOU CAN ENTER THE ENTIRE OVERIDE OR GENERATE AND EDIT
THE DEFAULT SQL STATEMENT.
LOOKUP OVERRIDE CAN INCLUDE WHERE CLAUSE ALSO...
HOW TO USE AN ORACLE SEQ GENERATOR IN A MAPPING?
WE HAVE TO WRITE THE STORED PROCEDURE

What is Constraint based loading in informatica?


By setting Constraint Based Loading property at session level
in Configaration tab we can load the data into parent and child
relational tables (primary foreign key).
Genarally What it do is it will load the data first in parent table
then it will load it in to child table.
what is target load plan?
here we can mention which target has to be load first.
What is use of Shortcuts in informatica?

If we copy source definaltions or target definations or mapplets


from Shared folder to any other folders that will become a
shortcut.

What are mapping parameters and variables in which


situation we can use it?
If we need to change certain attributes of a mapping after every time the session is run, it will be very difficult
to edit the mapping and then change the attribute. So we use mapping parameters and variables and define the
values in a parameter file. Then we could edit the parameter file to change the attribute values. This makes the
process simple.
Mapping parameter values remain constant. If we need to change the parameter value then we need to edit
the parameter file. But value of mapping variables can be changed by using variable function. If we need to
increment the attribute value by 1 after every session run then we can use mapping variables. In a mapping
parameter we need to manually edit the attribute value in the parameter file after every session run.

Why we need source qualifier?


Simply it performs select statement. Select statement fetches the data in the
form of row.
Source qualifier will select the data from the source table.It identifies the record
from the source.

What is Constraint based loading in informatica?


By setting Constraint Based Loading property at session level in
Configaration tab we can load the data into parent and child relational
tables (primary foreign key). Genarally What it do is it will load the
data first in parent table then it will load it in to child table.

Have you done any Performance tuning in informatica?


1) Yes, One of my mapping was taking 3-4 hours to process 40 millions
rows into staging table we dont have any transformation inside the
mapping its 1 to 1 mapping .Here nothing is there to optimize the
mapping so I created session partitions using key range on effective
date column. It improved performance lot, rather than 4 hours it was
running in 30 minutes for entire 40millions.Using partitions DTM will
creates multiple reader and writer threads.

2) There was one more scenario where I got very good performance in
the mapping level .Rather than using lookup transformation if we can
able to do outer join in the source qualifier query override this will give
you good performance if both lookup table and source were in the same
database. If lookup tables is huge volumes then creating cache is costly.
3) And also if we can able to optimize mapping using less no of
transformations always gives you good performance.
4) If any mapping taking long time to execute then first we need to look in
to source and target statistics in the monitor for the throughput and also
find out where exactly the bottle neck by looking busy percentage in the
session log will come to know which transformation taking more time ,if
your source query is the bottle neck then it will show in the end of the
session log as query issued to database that means there is a
performance issue in the source query.we need to tune the query using

What are the new features of Informatica 9.x in


developer level?
From a developer's perspective, some of the new features in Informatica 9.x are as follows:

Now Lookup can be configured as an active transformation - it can return multiple rows on
successful match

Now you can write SQL override on un-cached lookup also. Previously you could do it only
on cached lookup

You can control the size of your session log. In a real-time environment you can control the
session log file size or time

Database deadlock resilience feature - this will ensure that your session does not
immediately fail if it encounters any database deadlock, it will now retry the operation again. You
can configure number of retry attempts.

Difference between Informatica 7.x and 8.x


1--> Informatica PowerCenter 8 is service-oriented for modularity, scalability and flexibility. Version 7
doesn't have the web-based administration console.
2--> The Repository Service and Integration Service (as replacement for Rep Server and Informatica
Server) can be run on different computers in a network (so called nodes), even redundantly.
3--> Management is centralized, that means services can be started and stopped on nodes via a central
web interface.

4-->Client Tools access the repository via that centralized machine, resources are distributed
dynamically.
5-->It has a support for unstructured data which includes spreadsheets, email, Microsoft Word files,
presentations and .PDF documents. It provides high availability, seamless fail over, eliminating single
points of failure.
6-->Informatica has now added more tightly integrated data profiling, cleansing, and matching
capabilities.
7-->A new web based administrative console has been added.
8-->Java transformation is introduced.
9-->Ability to write a Custom Transformation in C++ or Java.
10-->Dynamic configuration of caches and partitioning
11-->New transforms like SQL transform, HTTP transform, Salesforce, etc. have been added in Version
8.

Difference between cached and un cached lookups?


For a cached lookup the entire rows (lookup table) will be put in the buffer, and compare
these rows with the incomming rows.
where as uncached lookup, for every input row the lookup will query the lookup table and
get the rows
WHAT IS WORKLET
A set of worlflow tasks is called worklet,
Workflow tasks means
1)timer2)decesion3)command4)eventwait5)eventrise6)mail etc......
But we r use diffrent situations by using this only

DTM Process
The Load Manager creates one DTM process for each
session
in the workflow. It performs the following tasks:
Reads session information from the repository.
Expands the server, session, and mapping variables and
parameters.
Creates the session log file.
Validates source and target code pages.

Verifies connection object permissions.


Runs pre-session shell commands, stored procedures and
SQL.
Creates and run mapping, reader, writer, and
transformation
threads to extract, transform, and load data.
Runs post-session stored procedures, SQL, and shell
commands.
Sends post-session email.
What are various types of Aggregation?
Various types of aggregation are SUM, AVG , COUNT, MAX, MIN, FIRST, LAST, MEDIAN,
PERCENTILE, STDDEV, and VARIANCE

What is code page in informatica


The code page in informatica is used to specify the character encoding. It is selected
based on the source data

Most commonly selected encoding systems are:


ASCII
UTF-8
UTF-32

Can u copy the session to a different folder or repository?


Yes. By using copy session wizard u can copy a session in a different folder or
repository.But that
target folder or repository should consists of mapping of that session.
If target folder or repository is not having the maping of copying session ,
u should have to copy that maping first before u copy the session
TYPES OF TRACING LEVELS?
When you configure a transformation, you can set the amount of detail the Integration
Service writes in the session log.
The session log tracing levels are:

1. Normal: Integration Service logs initialization and status information, errors


encountered, and skipped rows due to transformation row errors. Summarizes session
results, but not at the level of individual rows.

2. Terse: Integration Service logs initialization information and error messages and
notification of rejected data.

3. Verbose Initialization: In addition to normal tracing, Integration Service logs


additional initialization details, names of index and data files used, and detailed
transformation statistics.

4. Verbose Data: In addition to verbose initialization tracing, Integration Service logs


each row that passes into the mapping. Also notes where the Integration Service
truncates string data to fit the precision of a column and provides detailed transformation
statistics.
By default, the tracing level for every transformation is Normal.

how to drop index using stored procedure?


create procedure procedurename
begin
execute immediate 'drop index indexname'
end;

You might also like