how can one eliminate dupliacte data with out using
distinct option?
Answers were Sorted based on User's Feedback
Answer / sunil
By using aggrator select groupby port to eliminate
duplicates
| Is This Answer Correct ? | 31 Yes | 0 No |
Answer / prabha kamaraj
Using Group by command removes all duplicates records in a
table
| Is This Answer Correct ? | 5 Yes | 0 No |
Answer / sujatha
in the sq data has to be order by key column..then in the
expression trans we have to create a variable port to hold
the previous records key value...and one more variable port
to compare with the current records key value...if the prev
and curr are equal then the 2nd variable value will
get 'Y',if not equal its value will be 'N'.
Take a router ,process the rows which are having the 2nd
variable value as 'N' to the target....these are unique recs
| Is This Answer Correct ? | 4 Yes | 0 No |
Answer / ankit kansal
An Alternative way
Using Rank Transformation select group by all and only select only those rows whose indexes are one.
http://deepinopensource.blogspot.in/
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / arpan
U can use Sorter Transformation and there is an option as
Select Distinct.
| Is This Answer Correct ? | 3 Yes | 6 No |
Briefly explain your complete project(sales) flow, (ie. from source received from client, transformations, then despatch to end user) what are all the process. Kindly give step by step process.
(Integ) Start workflow: ERROR: User [practice] does not have sufficient privilege for this operation. how to solve this pls suggest me? THANKS in advance
How we can get unique records into one target table and duplicate records into another target table??
11 Answers Deloitte, IBM, iGate,
If we use only lookup transformation in a mapping ie, SourceQualifier-->Lookup --> Target. , here datas are taking very long time to load in target., so what are steps to improve the performance in that mapping???????
How union transformation is used?
following scenario i have 1000 record flatfile source i want ist row to ist target 2nd eow to 2nd target 3rd row to 3rd target how will u do?
What aer the out put files that the informatica server creates during the session running?
in which situations we go for pesistent cache in lookup and which situations go for shared lookup cache?
My source is delimited flat file Flat file data is H|Date D1|ravi|bangalore D2|raju|pune T|4 The data will be send to target if the fallowing two conditions satisfied 1.The first row Date column is equal to SYSDATE 2.Last record second port equal to number of records. How to achieve?
Can we write to_date(three arguments) ? Will it work ? Advance thanks
on a day i load 10 rows in my target and on nextday i get 10 more rows add to my target but out of 5 are update row how ican send them to target?how i can insert and update the records
I am hvaing SOURCE as 1000,null,null,null null,2000,null,null null,null,3000,null null,null,null,4000 Now i want the OUTPUT as 1000,2000,3000,4000 For more clarification i want to elimate nulls and want in a single line. Please help me out