Out of 4 mill records only 3 mill records are loaded to
target
and then job aborted. How to load only those 1 mill(not
loaded records) for next run.
This job is not sequential job, it is stand alone parallel
job.What are the possibilities available in datastage8.1?
Answers were Sorted based on User's Feedback
Answer / prasanna
Use already loaded records TGT as reference lookup,and use
lookup stage,with DROP option..and design the job.u ll get
req answer. refer
.
.
sour.......>lookup......>tgt
Is This Answer Correct ? | 7 Yes | 4 No |
Answer / nish
there are plenty of options available.
just carefully study the scenario.
Source: 4 Mil records (doesn't change)
Target: 3 Million already loaded
Option1:
you just need to identify those 1Million pool them and then load to target.
This is clearly a case for Change Data Capture (CDC) stage.
use the Source As Before Table and Target as After.
Write those 1 Mil records based on change_code() (Deleted) to a file.
Move the contents of this file to target.
Option 2: This scenario also hints at updating the target with merge stage.use the DROP option to gather into a file and then update the target.
Option 3: Look Up being a performance concern should be your alternative.
Is This Answer Correct ? | 2 Yes | 0 No |
Answer / venkatesh
first load the data using with surrogate stage
after that use filter stage and take that key of
surrogateand in lookup maintain the Drop option
Is This Answer Correct ? | 2 Yes | 8 No |
What is the difference between datastage and datastage tx?
How can we perform the 2nd time extraction of client database without accepting the data which is already loaded in first time extraction?
How many partitions does a node can have? Some body explain me clearly about the partitions and the nodes?
1.what is materialized data? 2.how to view the materialized data?
What is the difference between odbc and drs stage?
What are the various kinds of the hash file?
what is the diff between sequential file and fileset stages?
what is the difference between lookup stage reject link and merge stage reject link in datastage Parallel jobs? interm of output in Merge Reject link and Look Up Reject link ?
what is the use of DSattchJob?DetachJob? where can we find it?
what is mapping lookup
Explain usage analysis in datastage?
Why we use surrogate key?