How to identify bottlenecks in
sources,targets,mappings,workflow,system and how to
increase the performance?
Answers were Sorted based on User's Feedback
Answer / praveenkumar.b
Source:
Create the Filter transformation after all the Source
Qualifiers and make the filter condition FALSE so that the
data will not go beyond this trasformation. Then run the
session and find out the time taken from source. If you
feel there is some lack in performance, then suggest the
necessary index creation in Pre Session.
Note: If the source is File, then there is no possibility
of performance problme in source side
Target:
Delete the target table from the mapping and create the
same structure as a Flat file. Run the session and find out
the time taken to write the file. If you feel problem in
performance, then delete the INDEX of the table before
loading the data. In post Session, Create the same index
Note:If the target is File, then there is no possibility of
performance problme in target side
Mapping:
The below steps need to be consider
#1. Delete all the transformations and make it as single
pass through
#2. Avoid using more number of transformations
#3. If you want to use more filter transformation, then use
Router transformation instead of it
#4. Calculate the index and data cache properly for
Aggregator, Joiner, Ranker, Sorter if the Power center is
lower version. Advance version, Power center itself will
take care of this
#5. Always pass the sorted i/p's to Aggregator
#6. Use incremental aggregation
#7. Dont do complex calculation in Aggregator
transformation.
Session:
Increas the DTM buffer size
System:
#1. Increase the RAM capacity
#2. Avoid paging
Is This Answer Correct ? | 27 Yes | 1 No |
Answer / kalyan
Run ur session in Verbose Mode and check out the Busy
Percentage in the Log. IF its more at the Reader thread
than ur Source Query is the Bottleneck.Tune your SQ.
If its Writer thread, then you check with your target . May
be you need to drop and recreate the Indexes on the target
table.
If its the Transformation thread , then check with your
mapping logic. Concentrate More on Aggregator part..
Fine tune your logic. Don't drag the fields which are not
used to all the transformations. try to use as less
transformations as possible.
Cache your lookups .Whenever possible use the persistent
lookup concept.
This should help guys..
Is This Answer Correct ? | 17 Yes | 2 No |
Bottleneck in Informatica
Bottleneck in ETL Processing is the point by which the performance of the ETL Process is slowr.
When ETL Process is in progress first thing login to workflow monitor and observe performance statistic. I.e. observe processing rows per second. In SSIS and Datastage when you run the job you can see at every level how many rows per second is processed by the server.
Mostly bottleneck occurs at source qualifier during fetching data from source joiner aggregator Lookup Cache Building Session.
Removing bottleneck is performance tuning.
Is This Answer Correct ? | 3 Yes | 0 No |
Answer / srinu
identification of bottelnecks
target:configuring session to write to flatfiletarget
source:add filter t/r after sq t/t to false show that no
data is processed past the filter t/r,if it time takes to
run new session remains same to the original session there
is source bottel necks
mapping:add filter t/f before each target and set filter
condition to false,similar to source
session:use the collect performance data to identify the
session bottel necks
read from desk,write to disk counters other than zero,there
is bottelnecks
Is This Answer Correct ? | 3 Yes | 5 No |
what is unique constraint error ?
how to import multiple flat files in to single target where there is no common column in the flat files
what are the difference between Informatica 7.1 and 8.1?
i had a source containing business,sales,details column and i have to load it to a target but i have some bad records in it , but i have to load 70% of business records and 50% of sales and 95% of details records excluding bad records to achieve this what should be done and what all logic and tx should be used can anyone help? thanks in advance
how to delete duplicate records by using filter transfermation?
Without using Lookup & Sequence Generator, How to generate Sequence?
i having mapping emp---> sq_emp--->target using this mapping i have to load dept table values into target... (condition assume there is no primary and foreign key relation between emp(table x) and dept (table Y))
Please explain in detail with example about 1.Confirmed Dimension. 2.Junk Dimension. 3.Degenerated Dimension. 4.Slowly changing Dimensions
Scheduling properties,whats the default (sequential)
How to list Top 10 salary, without using Rank Transmission?
i have source file data empno,ename,sal 100,ram,1000 200,tyu,2000 300,gh,3000 out put as e empno,ename,sal 100,ram,null 200,tyu,1000 300,gh,2000 how to load it?
What is informatica? What exactly are the practical uses in real world