What are the mapings that we use for slowly changing
dimension table?
Answer Posted / swetha
Type1: Rows containing changes to existing dimensions are
updated in the target by overwriting the existing dimension.
In the Type 1 Dimension mapping, all rows contain current
dimension data.
Use the Type 1 Dimension mapping to update a slowly changing
dimension table when you do not need to keep any previous
versions of dimensions in the table.
Type 2: The Type 2 Dimension Data mapping inserts both new
and changed dimensions into the target. Changes are tracked
in the target table by versioning the primary key and
creating a version number for each dimension in the table.
Use the Type 2 Dimension/Version Data mapping to update a
slowly changing dimension table when you want to keep a full
history of dimension data in the table. Version numbers and
versioned primary keys track the order of changes to each
dimension.
Type 3: The Type 3 Dimension mapping filters source rows
based on user-defined comparisons and inserts only those
found to be new dimensions to the target. Rows containing
changes to existing dimensions are updated in the target.
When updating an existing dimension, the Informatica Server
saves existing data in different columns of the same row and
replaces the existing data with the updates.
| Is This Answer Correct ? | 3 Yes | 0 No |
Post New Answer View All Answers
Explain the types of lookup transformation?
I need the output file in the below format using Informatica. __0946684800__ 6 1 3 1 5 1 6 1 10 1 6 10 Pos State Time Split AgentID Reason Spaces Code in Code Spare state ______ ___ _____ ______ __________ ______ 1234 14 0 4321 3103 abcd 4321 15 5 123452 3150 Can someone help please?? We need the data in space-delimited, fixed ASCII format .
Explain pushdown optimization $pushdownconfig parameter - informatica
Define filter transformation?
Differentiate between source qualifier and filter transformation?
Some flat files are there, out of these having some duplicate. How do you eliminate duplicate files while loading into targets?
During the running session, output files are created by informatica server. Enlist few of them.
What differs when we choose the sorted input for aggregator transformation?
Which version configuration tool used in ur project?
How will the document be delivered to me?
IN SCD1, insource we have 10 billion records and in the first day its uploaded successfully and in the second day its taking time to upload because some records it might get update or insert new records. As a developer what will be the better solution for this??
What is standalone command task?
I am not able to connect to the domain with the client although all services and databases are up and there is no network issue?
Explain Dataware house architecture .how data flow from intial to end?
What all join we can perform in LOOKUP transformation?