why we are using level option in normalizer transformation
Answer Posted / lokesh
If you want to see result in quarterly based with all other information like location, branch then we can go for normalizer transformation.
Your normalizer transformation looks like this for above example:
Field Level Occurs Description
Location 1 0
Branch 2 2 In each record we can see branch detail occurs two times
result 2 4 For each branch 4 times we can see quarter result, so in a record 4(occurs) X 2 (level) = 8 times
It automatically create input/output port depends on your level and occurs. Also generates some output key ports. You can see more detail on key port in help material.
Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers
Workflow is long running due to long running sql query so when we refer the query plan it tells the issue is due to partition of the db table. How to handle this?
Describe the scenarios where we go for joiner transformation instead of source qualifier transformation?
Clarify the utilization of aggregator cache record?
If informatica has its scheduler why using third party scheduler?
How do you load first and last records into target table?
Whats the difference between informatica powercenter server, repositoryserver and repository?
What does reusable transformation mean?
write a query to retrieve the latest records from the table sorted by version(scd)
Please tell me which institute is the best to study Informatica and asp.net in chennai Also please send me the latest interview questions in asp.net,c# and sql server to my id ramtryin@gmail.com
how lookup transformation is made active in new versions... When to use connected and when to use unconnected lookup and why? which is good for session performance. How to make lookup persistent and how to remove stale data from that lookup. how commit works - when we stop or abort data. Explain in both cases. What is factless fact table and have you ever used it in real time scenarios.
How to join three sources using joiner? Explain though mapping flow.
If the source has duplicate records as id and name columns, values: 1 a, 1 b, 1 c, 2 a, 2 b, the target should be loaded as 1 a+b+c or 1 a||b||c, what transformations should be used for this?
What are the various test procedures used to check whether the data is loaded in the backend, performance of the mapping, and quality of the data loaded in informatica?
Clarify the aggregator change?
write sql query vertical to horiozontal following table id name 1 100 2 dinesh 3 india 1 101 2 suresh 3 india 1 103 2 prakesh 3 usa i want output like id name country 100 dinesh india 101 suresh india 103 prakesh usa