Converting Rows to columns
I have Relational source like his.
JAN FEB MAR APR
100 200 300 400
500 600 700 800
900 100 200 300
I need to convert these rows into columns to the targe.
MONTH TOTAL
JAN 1500
FEB 900
MAR 1200
APR 1500
Please experts help me
Answers were Sorted based on User's Feedback
Answer / raj
sourcequalifier....>normalizer.....>expression...>aggrgater....>target
| Is This Answer Correct ? | 8 Yes | 1 No |
Answer / usha
by using normalizer transformation we can convert rows to
columns.
s-->SQ--->NRM-->Target
| Is This Answer Correct ? | 2 Yes | 0 No |
Answer / dev
hi Raj, plz explain, what we need to do in expression
transformation...
| Is This Answer Correct ? | 0 Yes | 0 No |
SQ->EXP->NRM->TGT
EXP :-- CREATE 4 OUTPUT PORTS
PORT1= 'JAN'
PORT2='FEB'
PORT3='MAR'
PORT4='APRIL'
AGGR :-- DRAG ALL PORTS FROM EXP AND CREATE 4 OUTPUT PORT
PORT5= SUM(JAN)
PORT6=SUM(FEB)
PORT7=SUM(MAR)
PORT8=SUM(APRIL)
NRM : CREATE 2 PORTS
MONTHS- OCCUR(4)- STRING
TOTAL- OCCUR(4)- NUMBER
CONNECT NRM OUTPUT PORT TO TGT
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / shankar
source qualifier > expression <for best practice> > aggregator > normalizer . target.
here in this the SQ will be for relational & flat files, then we use expression for best practice, in this we take the ports which we need into target also we can rename the ports for our convenience just to avoid confusion.
Then for calculating totals we use aggregator.
to convert the rows to col we use the normalizer. This is to my knowledge. Thank you if wrong pls post.
| Is This Answer Correct ? | 0 Yes | 2 No |
In a table there are 1 million records there in which 3 records are duplicate how will you find out those 3 records?
Can we write to_date(three arguments) ? Will it work ? Advance thanks
one source to target loading insert new rows and update existing rows?
Two workflows are running at same time first workflow is succeeded but second workflow is failed but there is no dependencies?
what is curr val use for in sequence generator?
How to fasten loading of 100 million distinct records in informatica? (Initially they are loaded into target without using any transformation, taking 2 hours)
Stage is which type of load is it a full load or incremental load Please let me know in detail with best example
How to partition the Session?
what are the different types of transformation available in informatica. And what are the mostly used ones among them?
why we use materialized view over view?
major difference between normal loading and bulk loading?
how much memory (size) occupied by a session at runtime