from Source 100 rows are coming, on target there are 5 m
rows which options is better to match data
1. Joiner
2 No cache
3. Static
4. Dynamic
Answers were Sorted based on User's Feedback
Answer / coolboy
Hi,
Here we will use joiner for better performance.We will join
the two sources making source table as master source.so
only 100 comparisons will be done.So it will be very faster.
Whereas in static and dynamic we have to look up on the
target which is very large 5m rows.So caching will take
more time.
Is This Answer Correct ? | 27 Yes | 3 No |
Answer / arnab
Lookup with No cache will make it faster , personal opinion
though , there are other ways do it also , if you have a
common key between the source and the target and they are
on the same database , it would be easier and faster to do
a join in the source qualifier itself and let the database
handle the query rather than Informatica server
Is This Answer Correct ? | 6 Yes | 3 No |
Answer / vijaykumar
Answer 7# is wrong because here we have only 1 source. It is
correct in case of 2 sources.
Is This Answer Correct ? | 2 Yes | 0 No |
Answer / jaybrata
Unless you want to join in Database, the only option considering performance is - USE NO CACHE.
BUT, you have to have proper index defined on your target database table, this is crucial as otherwise there will be 100 full table scans on 5 million records.
Static Cache is the worst option as it will always cache 5 million records just to compare 100 records.
Dynamic cache is not an option assuming 100 source records do not contain duplicate.
Informatica side joining is also not good, better to join in DB side instead (To know why and to check the comparison between Infa side and DB side joins, check www.dwbiconcepts.com)
Is This Answer Correct ? | 1 Yes | 0 No |
Answer / jaspreet banga
Its better to use connected dynamic non-persistent lookup
transformation, to lookup into the traget and if record
exist , update it else insert the new using unconnected
lookup tranmsformation..
Is This Answer Correct ? | 1 Yes | 0 No |
Answer / ajay bhosal
Uncached lookup can give you good performance here but only when your target is properly indexed. Moreover, it's not a good idea to hit the database 100 times and scan huge data volume in the database.
Similar or better performance can be achieved by using Joiner with Sorted Input, but ensure that 100 rows are in master pipeline and the detail pipeline is sorted.
There is no question of using cached lookup.
Answer will be Joiner.
Is This Answer Correct ? | 1 Yes | 0 No |
Answer / srik
5m => 5 million recors.
1) Since target side records are very huge compared to
the source, caching is not preffred. so No caching.
2) Again with no caching we can accomplsh this in 2 ways
a) By Joiner
b) Unconnected lookup
but as unconnected lookup is not in answers is in the
option, Joiner is the answer
Is This Answer Correct ? | 2 Yes | 2 No |
Answer / prakash
USE NOCACHE,IT WILL GIVE BETTER PERFORMANCE RATHER THAN
JOINER
Is This Answer Correct ? | 1 Yes | 1 No |
what is the architecture of any Data warehousing project?
major difference between normal loading and bulk loading?
How to create a mapping ? id date 101 2/4/2008 101 4/4/2008 102 6/4/2008 102 4/4/2008 103 4/4/2008 104 8/4/2008 O/P - shud have only one id with the min(date) How to create a mapping for this
What are the rank caches?
what is mapping optimization? wat are the techniques for tat
What is the difference between a repository server and a powerhouse?
suppose my source is 101 a 1o1 b 101 c 101 d i want target like that 101 abcd how will u achive this please give me the answer
WHAT IS THE NAME OF THAT PORT IN DYNAMIC CACHE WHICH IS USED FOR INSERT , UPDATE OPRATION ?
Define update strategy?
i have two sources both is oracle database . one is coming from USA database and another one is coming from U.k database .i am using source qualifier transformation to join this.as that time in session level source properties which path i need to give to retrieve that data. thanks bala 09619894486
diffrence between command task and control task
Converting Rows to columns I have Relational source like his. JAN FEB MAR APR 100 200 300 400 500 600 700 800 900 100 200 300 I need to convert these rows into columns to the targe. MONTH TOTAL JAN 1500 FEB 900 MAR 1200 APR 1500 Please experts help me