how will you remove the duplicate records from flat file
without using sorter?
Answers were Sorted based on User's Feedback
Answer / rkumar.etl
use aggregator transformation and group by all ports and
create one port for checking count...and pass the results
accordingly to target tables.
| Is This Answer Correct ? | 24 Yes | 3 No |
Answer / sathish
use aggregater T/R after source qualifier select all ports in group by clause it will removes the duplicates
| Is This Answer Correct ? | 5 Yes | 2 No |
Answer / ajit kumar pradhan
if it is a relational source then click on source Qualifier
properties then check Distinct option.
if it is a flat file then it can't work .......
click on target defination select target menu and click
generate/execute then it come out one dialog box come out
then check primary key check box..it avoid duplicale row...
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / sathish
put primary key to target it will reject duplicate rows
| Is This Answer Correct ? | 1 Yes | 4 No |
Answer / amedela chandra sekhar
SD---S.Q----EXP t/r------TGT.
add two port in exp t/r.
one is variable port v_seq expression:seq+1
one output port o_seq expression:v_seq
required ports connect to target.
| Is This Answer Correct ? | 1 Yes | 4 No |
We can use the option 'SELECT DISTINCT' in the source
qualifier.
Anybody please tell me whether this works or not.
| Is This Answer Correct ? | 5 Yes | 23 No |
What is the difference between a repository server and a powerhouse?
which transformation should we use to get the 5th rank member from a table in informatica?can we achieve this in sql?
What is a filter transformation?
Consider two cases: (1) Power Center Server and Client on the same machine (2) Power Center Sever and Client on the different machines what is the basic difference in these two setups and which is recommended?
Target is zero file and how can you send the email?
we have tables like c1 a c2 b c3 c c4 x c5 y and i need output like abcx in a single row and abcy in a single row? how do u do this?
2 Answers Patni, Tech Mahindra,
a soure table has 100 records how to load into 3 target tables in a cumulative order source target1 target2 target3 --------- ------------------------------------- 1-100 1 2 3 4 5 6 - - - 98 99 100
One of the optimizing technique to improve the session performance is push down optimization,by using push down optimization we push as much as transformation logic to source/target database,but this degrades the d/b performance,how to overcome this?
What is the benefit of session partitioning?
What is enterprise data warehousing?
What is dimensional table?
In a simple mapping when i use aggregate t/r with out any condition or group by, i am getting output as only last record in the table . can any one explain it please.........?