How to load the data into target table, which not have
Primary/Surrogate key (Duplicate records not acceptable)
Answers were Sorted based on User's Feedback
Answer / babu
u can write the filter condition in filter t/r.
in aggregator t/r select all ports are group by and add one
out put port,that is out--count(*),
in filter t/r u cam place the condition is out=1;
mapping:
src->sq->agg(t/r)->filtr or router t/r->trg
Is This Answer Correct ? | 4 Yes | 0 No |
Answer / ramesh
this can achieved using Update statergy
we have option in UPD or Target update override
Is This Answer Correct ? | 3 Yes | 1 No |
Answer / santosh kumar sarangi
1. If source is relational then in SQ you can check the
distinct or in mapping we can use distinct in sorter t/r.
2.if flat file then distinct in sorter t/r.
It will not allow nay duplicate data, so target will get
only distinct data.
Let me know if any thing wrong.
Thanks & regards
Santosh Kumar Sarangi
Is This Answer Correct ? | 1 Yes | 0 No |
Answer / upendra
sq--> srtr-->aggr-->tgt
note:
srtr->properties->
select distinct and connect to aggr tr
aggr-->group by key column
aggr->properties->select sorted input
and connect to tgt
Is This Answer Correct ? | 0 Yes | 0 No |
Answer / moorthy g [ibm]
1. Target Update Override
2. Select Unique Value (means)
Exp:
out_flag: decode( true, pre_no=no, 'Y', 'N')
pre_no = no
Router:
Out_flag: Y --- Unique Value
Default --- Duplicate Value
Is This Answer Correct ? | 0 Yes | 1 No |
wt is Dynamic lookup Transformation? when we use?how we use?
two tables from two different databases r there . both having same structure but different data . how to compare these two tables ?
Where should yoU place the flat file to import the flat file defintion to the designer?
Data transformed successfully from Source table to target table. Now how you will ensure that the data in Target table is proper. I answered will verify one or two records and check Then question was that the Development is doing (1-2 record verification)but as a tester you have to verify the complete data, how you will do?? Please answer
if i have one mapping and we already tuned that mapping for performance,everything is fine and loading will take 1 hr,so without doing any change in mapping how could we reduce the loading time from 1hr to 1/2 hr.
wf dont have integration severances how you can run?
in which situations do u go for snowflake schema ?
Want to know about Training Centres for Informatica, Cognos and ETL Softwares in Mumbai, India.
What if the source is a flat-file? Then how can we remove the duplicates from flat file source?
What is Micro Strategy? Why is it used for?
what is the difference between stop and abort?
without table how to come first record only in oracle?