What is the default execution order of the sequential file?

Answer Posted / shar

It behaves different w r t reading capability.
i.e.,
when it is reading a single file; it does sequential mode of
reading by default and when it is reading multiple files it
does parallel mode of reading by default.

And if we need to improve performance of sequential file
stage when reading a single large file then we have two
options:
1. Read from Multiple Nodes
2. No of readers per node.

1st one is applicable only when flat file is Fixed length
file and not applicable for delimited formatted files.

The latter one is applicable for delimited files by
mentioning no of nodes then it performs parallel mode of
execution by reading single file in partitions improving the
performance.

Thats it.

Is This Answer Correct ?    4 Yes 0 No



Post New Answer       View All Answers


Please Help Members By Posting Answers For Below Questions

What is active and passive stage?

887


Why we use surrogate key?

701


What is meta stage?

584


what is the use of skid in reporting?

1961


What is apt_config in datastage?

710






What are routines in datastage? Enlist various types of routines.

631


How do you import and export the datastage jobs?

741


Source has 2 columns: USA,NewYork INDIA,MUMBAI INDIA,DELHI UDS,CHICAGO INDIA,PUNE i want data in target like below: INDIA,MUMBAI1 INDIA,DELHI2 INDIA,PUNE3 USA,NEWYORK1 USA,CHICAGO2

356


1)what is the size of Fact table and dimension table? 2)how to find the size of Fact table and dimension table? 3)how to implement the surrogate key in transform stage? 4)write the configuration file path? 5)how many types of datasets explain? 6)diff b/w developed projects and migration projects? 7)how to delete the header and footer file of the sequencer file? 8)how can u call the parameters in DS in unix environment? 9) how much data ur getting daily ? 10)

2327


create a job that splits the data in the Jobs.txt file into four output files. You will direct the data to the different output files using constraints. • Job name: JobLevels • Source file: Jobs.txt • Target file 1: LowLevelJobs.txt − min_lvl between 0 and 25 inclusive. − Same column types and headings as Jobs.txt. − Include column names in the first line of the output file. − Job description column should be preceded by the string “Job Title:” and embedded within square brackets. For example, if the job description is “Designer”, the derived value is: “Job Title: [Designer]”. • Target file 2: MidLevelJobs.txt − min_lvl between 26 and 100 inclusive. − Same format and derivations as Target file 1. • Target file 3: HighLevelJobs.txt − min_lvl between 101 and 500 inclusive. − Same format and derivations as Target file 1. • Rejects file: JobRejects.txt − min_lvl is out of range, i.e., below 0 or above 500. − This file has only two columns: job_id and reject_desc. − reject_desc is a variable-length text field, maximum length 100. It should contain a string of the form: “Level out of range: ”, where is the value in the min_lvl field. My Question is how do you write the stage variable for reject rows.

2201


What are the main differences you have observed between 7.x and 8.x version of datastage?

626


Difference between in process and inter process?

705


What is a datastage job?

625


How to read the length of word in unix?

888


How we can covert server job to a parallel job?

601