Answer Posted / guru avula
1.In a project source systems are differnet and also all
source systems are not availble in same time.for eg 1
source is availble at 1AM and 2n one at 2 AM etc
But our schedule jobs run at different time ,so we have to
pick the data from source system based on source availble
time and put into staging area.
2.All source systems table formats and column formats are
diffrent ,so we have to sync all .
For above two reasons we go for staging area
| Is This Answer Correct ? | 5 Yes | 1 No |
Post New Answer View All Answers
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
How do you start developing a datastage project?
what is stage is used for below Input columns: dept|mgr|employee|salary Output columns: mgr|count of employee per mgr|avg salary per dept note: each dept has one mgr and each mgr has many employees
1.new record it will insert but changes of natural key is not present in taget i want to update (here key is composite natural key )can any one help this to explan how to do
Define data aggregation?
What is usage analysis in datastage?
Can anyone tell me a difficult situation who have handled while creating Datastage jobs?
DB2 connector> transformer > sequential file Data will be exported into a csv format in a sequential file. This file will be send in a email using a sequence job. Problem here is, how to avoid sending a blank csv file? When I ran the job there are chances that it might return zero records but in the sequence job csv file is going blank. how can I avoid this? thanks
Where the datastage stored his repository?
What all are the different way to run a job?
What is aggtorec restructure operator?
Is it possible to implement parallelism in Mainframe Jobs ? If Yes how ? If no why ?
Differentiate between validated and Compiled in the Datastage?
Can you highlight the main features of ibm infosphere information server?
How to clean the datastage repository?