How can yoU improve session performance in aggregator
transformation?
Answers were Sorted based on User's Feedback
Answer / haritha
use sorted input option to decrease the use of aggregator
cache.
use filter transformation before aggregator transformation
to to rduce unnecessary aggregation.
Limit the number of connected input/output or output ports
to reduce the amount of data the Aggregator transformation
stores in the data cache.
| Is This Answer Correct ? | 39 Yes | 1 No |
Answer / swetha
Use sorted input.
The aggregator stores data in the aggregate cache until it
completes aggregate calculations. When u run a session that
uses an aggregator transformation, the informatica server
creates index and data caches in memory to process the
transformation. If the informatica server requires more
space, it stores overflow values in cache files.
| Is This Answer Correct ? | 13 Yes | 5 No |
Answer / dilip
if you sort the input yet do not select the sorted input
option, the aggregator treats the data as unsorted data &
performs the task hence its cache size increases. the
performance is degraded despite using a sorter.
| Is This Answer Correct ? | 8 Yes | 2 No |
Answer / chinna
what u guys said is so far ok
but what happen if u have't check the option sorted in
input in aggregator transformation but u actully sorted
before with sorter transformation so what will happen in
agg by performance wise?
| Is This Answer Correct ? | 3 Yes | 0 No |
Answer / srikanth
hi haritha
u r correct
For using sorted input option, the order of group by ports
in the agg and sorter transfomratoins should be same as
well as the number of ports also same.
we can filter out data early in the morning by using sq
trans rather than using filter before agg hence reducing
unwanted data in teh mapping.
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / sasmita panda
session performance in aggregator T/R can be increased by
aggregation cache i.e through increamental aggregation.
Aggregation cache has tow caches i)index ii) data.
For first run it performs aggregation calculation in
aggregation cache.Index cache contains group by port
values.Data cache will contain if any variable ports are
created and output port on which we want to implement
aggregate function.
Rowid of last record which is participated in Agg. T/R will
be saved in repository.
During second run it finds the rowid of last record then it
performs increamental aggregation.The integration service
will check the historical information in the index cache
file for corresponding group.If it finds it performs
aggregation operation increamentally. If not finds then it
will adda new port for aggreagtion.
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / pg
Chinna,
If u have't check the option sorted in
input in aggregator transformation but u actully sorted
before with sorter transformation in that case your session
will fail. You must check the option sorted in
input in aggregator transformation if you are using sorted
input to the aggregator .
Plz let do me know if you still have doubt.
pramodgup@gmail.com
| Is This Answer Correct ? | 6 Yes | 16 No |
Answer / cherry
hi PG
u r absolutely right.
session will fail.
u have to use the check option for sorted input.
| Is This Answer Correct ? | 4 Yes | 14 No |
i have source data like id name sal 1 a 65020 2 b 78250 3 c 58264 but how to get target like this id name sal 1 a Sixty five thousand twenty 2 b Seventy thousand two fify 3 c Fifty eight thousand two sixty four
What is Target Update Override? What is the Use ?
hi i want to install latest informatica software in my pc..can anybody help me where should i get the infomatica software..plz help me
What is partioning?how many types of partinings are there
i have a wf like wf-->s1-->s2-->s3-->s4 first start s1 later s2 later s3 here my session s3 have to run 3 times later start s4?
I am not able to connect to the domain with the client although all services and databases are up and there is no network issue?
What is a joiner transformation and why it is an active one?
What is meant by incremental aggregation?
How can the following be achieved in 1 single Informatica Mapping. * If the Header table record has error value(NULL) then those records and the corresponding child records in the SUBHEAD and DETAIL tables should also not be loaded into the target(TARGET1,TARGET 2 or TARGET3). * If the HEADER table record is valid, but the SUBHEAD or DETAIL table record has an error value (NULL) then the no data should be loaded into the target TARGET1,TARGET 2 or TARGET3. * If the HEADER table record is valid and the SUBHEAD or DETAIL table record also has valid records only then the data should be loaded into the target TARGET1,TARGET 2 and TARGET3. =================================================== HEADER COL1 COL2 COL3 COL5 COL6 1 ABC NULL NULL CITY1 2 XYZ 456 TUBE CITY2 3 GTD 564 PIN CITY3 SUBHEAD COL1 COL2 COL3 COL5 COL6 1 1001 VAL3 748 543 1 1002 VAL4 33 22 1 1003 VAL6 23 11 2 2001 AAP1 334 443 2 2002 AAP2 44 22 3 3001 RAD2 NULL 33 3 3002 RAD3 NULL 234 3 3003 RAD4 83 31 DETAIL COL1 COL2 COL3 COL5 COL6 1 D001 TXX2 748 543 1 D002 TXX3 33 22 1 D003 TXX4 23 11 2 D001 PXX2 56 224 2 D002 PXX3 666 332 ======================================================== TARGET1 2 XYZ 456 TUBE CITY2 TARGET2 2 2001 AAP1 334 443 2 2002 AAP2 44 22 TARGET3 2 D001 PXX2 56 224 2 D002 PXX3 666 332
Explain about Recovering sessions?
How to Explain My project flow,In TCS ased me this question. In which way i can start my project flow and they asked how many fact tables and dimensional tables u used. Can any one Explain Briefly for this question and project architecture also.. please..
3 Answers CTS, HCL, Infosys, TCS,
What is workflow? What are the components of workflow manager?