What is the default execution order of the sequential file?
Answer Posted / shar
It behaves different w r t reading capability.
i.e.,
when it is reading a single file; it does sequential mode of
reading by default and when it is reading multiple files it
does parallel mode of reading by default.
And if we need to improve performance of sequential file
stage when reading a single large file then we have two
options:
1. Read from Multiple Nodes
2. No of readers per node.
1st one is applicable only when flat file is Fixed length
file and not applicable for delimited formatted files.
The latter one is applicable for delimited files by
mentioning no of nodes then it performs parallel mode of
execution by reading single file in partitions improving the
performance.
Thats it.
Is This Answer Correct ? | 4 Yes | 0 No |
Post New Answer View All Answers
What is active and passive stage?
Why we use surrogate key?
What is meta stage?
what is the use of skid in reporting?
What is apt_config in datastage?
What are routines in datastage? Enlist various types of routines.
How do you import and export the datastage jobs?
Source has 2 columns: USA,NewYork INDIA,MUMBAI INDIA,DELHI UDS,CHICAGO INDIA,PUNE i want data in target like below: INDIA,MUMBAI1 INDIA,DELHI2 INDIA,PUNE3 USA,NEWYORK1 USA,CHICAGO2
1)what is the size of Fact table and dimension table? 2)how to find the size of Fact table and dimension table? 3)how to implement the surrogate key in transform stage? 4)write the configuration file path? 5)how many types of datasets explain? 6)diff b/w developed projects and migration projects? 7)how to delete the header and footer file of the sequencer file? 8)how can u call the parameters in DS in unix environment? 9) how much data ur getting daily ? 10)
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
What are the main differences you have observed between 7.x and 8.x version of datastage?
Difference between in process and inter process?
What is a datastage job?
How to read the length of word in unix?
How we can covert server job to a parallel job?