Please explain me the difference between 3 types of slowly
changing dimension in datawarehousing?
Answer Posted / rameshgoud
scd1-> with this process we can maintain only updated data,
for ex- if a record inserted in source then the same record
should be inserted in targer, if a record updated in the
source then same update should process in target, so here
we cant maintain the history
scd2->with this we can maintain current data and complete
historical data by adding the start_date and end_date of
the records in the target table. If a records get updated
in the source same record will insert in the target as new
record and the old record is updated with end date as
todays date. Like wise we there will no be any deletion of
records, so we can maintain the compelte history here.
scd3-> using this we can maintain current and recent
historical data only
for every source possible changing column we need two
target columns as NEW_COLUMN indicates current data and
OLD_COLUMN indicates recent historical data
when a new record getting loaded source data is always
loded in NEW_COLUMN in target
when a record is midified target NEW_COLUMN is updated in
target OLD_COLUMN and source data is updated in target
NEW_COLUMN.
| Is This Answer Correct ? | 17 Yes | 1 No |
Post New Answer View All Answers
What are the different common services in datastage?
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
What is the use of hoursfromtime() function in transformer stage in datastage?
Can you explain how could anyone drop the index before loading the data in target in datastage?
What are the different type of jobs in datastage?
What is the sortmerge collector?
Define meta stage?
How do you reject records in a transformer?
Explain entity, attribute and relationship in datastage?
What is the difference between datastage and informatica?
What is the purpose of pivot stage?
im new to this tool im now at project plz tell me step by step process how to design plz help me i wnt to go with exp for job plz give me d proper design and explination
What is usage analysis in datastage?
Describe the main features of datastage?
How you can fix the truncated data error in datastage?