what is sparse lookup?
Answers were Sorted based on User's Feedback
Answer / ramya
use sparse lookup when reference database is huge i.e ratio
of input data and reference data is 1:100 or more
| Is This Answer Correct ? | 37 Yes | 7 No |
Answer / venugopal
When the reference data is more and source data is very
very less at that time we are using the Sparce lookup i.e
the ratio is 1:100
| Is This Answer Correct ? | 23 Yes | 5 No |
Answer / prasad
whenever the reference data is very huge than the primary
data then it will be problem for usig lookup stage,b'se the
data stored in buffer in lookup stage.using sparse lookup we
can perform the lookup in database level not in the
buffer,so the performance can be automatically improved...
| Is This Answer Correct ? | 5 Yes | 1 No |
Answer / chenna
sparse lookup has more memory so that if reference table
contains huge amount of data perfer to use sparse lookup
| Is This Answer Correct ? | 3 Yes | 4 No |
Answer / ananda
Join: When volume of data is more than 10 million in both
source and reference link.
Sparse Lookup: When reference data is > 500 million and
source data is <10 million.
| Is This Answer Correct ? | 2 Yes | 8 No |
Answer / vin
If reference data is huge then we will go for join stage.
Then when we will use join and sparce is useful?
| Is This Answer Correct ? | 1 Yes | 15 No |
1.How to read multiple file from sequential stage. 2.If a file doesn't arrive or doesn't exists in sequential stage how do u handle this. 3.What do you do before taking data from source to staging area. 4. I have a remove duplicate stage and transformer stage.what will u do for optimizeing the performance of the job.
What is use Array size in datastage
Differentiate between hash file and sequential file?
what is the custome stage in datastage? how can we impliment that one? plz tell me
Can you explain engine tier in information server?
What is exact difference between Parallel Jobs and server Jobs..
can any one tell me how to install datastage 8.1 in windows xp with wizard
what are fact tables and dimension tables? give example assuming one table.
which is more expensive hash or modulus partitioning? when do you use modulus partitioning?
if we using two sources having same meta data and how to check the data in two sources is same or not? and if the data is not same i want to abort the job ?how we can do this?
Explain Quality stage?
1)Source file contains one record, I want 100 records in target file.