I have load a Dataset in UAT with 2 Node configuration, imported the job into PROD environment which is 4 node configuration and using this DataSet as SRC to other job. will the job run fine or give any errors? If job runs fine, on how many nodes? 2 nodes or 4 nodes?
Answer Posted / ankit gosain
Hi,
Job will run smoothly but the dataset will run on 2 node
configuration.
This is because whenever you create a dataset, 4 componets
are created which are Discriptor file, Data file, Header
File and Control file. Now, when that dataset was created
with 2 node configuration, it captured the configuration
file into it's Discriptor file and now if you'll use that
dataset in some other environment with 4 node
configuration, it won't affect anything, since the
discriptor have the configuration of nodes when it was
creadted.
You can cross check it by using the environment variable
APT_DUMP_SCORE.
If you have further doubts or query, you can mail me at
ankitgosain@gmail.com
Cheers,
Ankit
| Is This Answer Correct ? | 6 Yes | 1 No |
Post New Answer View All Answers
hi.... am facing typical problem in every interview " I need some critical scenarios faced in real time" plz help me guys
what is ds administrator used for?
Define ds designer?
Define data aggregation?
What is a folder? Difference types of stages?
How complex jobs are implemented in datstage to improve performance?
What are data elements?
How many Key we can define in remove duplicate stage?
What could be a data source system?
What all the types of jobs you developed?
Define meta stage?
What are the difference types of stages?
Define oconv () and iconv () functions in datastage?
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
what is the use of materialized view?