I HAVE A SOURCE FILE CONTAINING
1|A,1|B,1|C,1|D,2|A,2|B,3|A,3|B
AND IN TARGET I SHOULD GET LIKE
1|A+B+C+D
2|A+B
3|A+B
WHICH TRANSFORMATION I SHOULD USE
Answer Posted / nitin
1) Src-->SQ->Normalizer-->Target
The first pipeline breaks the rows and convert to columns and then write to target file.
1|A
1|B
1|C
1|D
2|A
2|B
3|A
3|B
2)Src->SQ->Exp->Target
The second pipeline receives the above output with '|' as delimiter and passes to expression.
In the expression transformation concatenate the second column with + for the same first column and then write to target with | as delimiter
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers
What are the tasks that source qualifier perform?
What is primary and backup node?
How to generate or load values in to the target table based on a column value using informatica etl tool.
What are the performance considerations when working with aggregator transformation?
How to load the name of the current processing flat file along with the data into the target using informatica mapping?
Some flat files are there, out of these having some duplicate. How do you eliminate duplicate files while loading into targets?
What is joiner transformation?
Separate from an archive server and a powerhouse?
In development project what is the process to follow for an etl developer from day1
Explain target update override in informatica
Define filter transformation?
Hi, In source I have records like this No name address 10 manoj mum 10 manoj dilhi 20 kumar usa 20 kumar Tokyo I want records in target like shown below No name addr1 addr2 10 manoj mum dilhi 20 kumar usa Tokyo If it is reverse we can do this by using Normalizer transformation by setting occurance as 2. Somebody will say use denoralization technique. But as of my knowledge I couldn’t find any denormalization technique. Is there any concept like that? I tryid this seriously but I could find any idea to implement this. Can any one please help me ? Advance Thanks
What is a joiner transformation?
Where are the source flat files kept before running the session?
One of the optimizing technique to improve the session performance is push down optimization,by using push down optimization we push as much as transformation logic to source/target database,but this degrades the d/b performance,how to overcome this?