how does work server jobs?



how does work server jobs?..

Answer / archana

When we have lots of functionality to implement for lower volume and hardware is less and ease of implementation we can go for Server jobs.
We can improve the performance of server job by enabling inter process row buffering. This helps stages to exchange data as soon as it is available in the link. IPC stage also helps passive stage to read data from another as soon as data is available.
The choice of server or parallel depends upon time to implement, functionality and cost.
Server jobs do not run on multiple node.
server job runs on windows platform usually
server job runs on one node
Server jobs process in sequence one stage after other
Server jobs doesn't support partition parallelism
server jobs the transformer is compiled in Basic language.

Is This Answer Correct ?    1 Yes 0 No

Post New Answer

More Data Stage Interview Questions

What are the environmental settings for data stage,while working on parellel jobs?

2 Answers   Satyam,


A flat file contains 200 records. I want to load first 50 records at first time running the job, second 50 records at second time running and so on, how u can develop this job?

1 Answers   TCS,


What are the primary usages of datastage tool?

0 Answers  


Hi, i did what you mentioned in the answer, i.e. source- >Transformer -> 3 datasets. Iam able to see the data in datasets but its not sort order... Can you tell how sort the data?? i also checked Hash partition with performsort.

1 Answers   CGI,


Hi Can any one help regarding below INPUT NAME LOC Ram hyd Ram ban Raj chn Raj Pun Sam del OUPUT NAME LOC Ram Hyd ban Raj chn pun sam del

3 Answers   HCL,






create a job that splits the data in the Jobs.txt file into four output files. You will direct the data to the different output files using constraints. • Job name: JobLevels • Source file: Jobs.txt • Target file 1: LowLevelJobs.txt &#8722; min_lvl between 0 and 25 inclusive. &#8722; Same column types and headings as Jobs.txt. &#8722; Include column names in the first line of the output file. &#8722; Job description column should be preceded by the string “Job Title:” and embedded within square brackets. For example, if the job description is “Designer”, the derived value is: “Job Title: [Designer]”. • Target file 2: MidLevelJobs.txt &#8722; min_lvl between 26 and 100 inclusive. &#8722; Same format and derivations as Target file 1. • Target file 3: HighLevelJobs.txt &#8722; min_lvl between 101 and 500 inclusive. &#8722; Same format and derivations as Target file 1. • Rejects file: JobRejects.txt &#8722; min_lvl is out of range, i.e., below 0 or above 500. &#8722; This file has only two columns: job_id and reject_desc. &#8722; reject_desc is a variable-length text field, maximum length 100. It should contain a string of the form: “Level out of range: <min_lvl>”, where <min_lvl> is the value in the min_lvl field. My Question is how do you write the stage variable for reject rows.

0 Answers   HCL,


How and where you used hash file?

0 Answers  


what is meta data? Explain? Where it is used?

9 Answers  


Why we use parameters instead of hard code in datastage.

2 Answers   IBM,


how to define satge variables in transformer stage

2 Answers  


if a column contains data like abc,aaa,xyz,pwe,xok,abc,xyz,abc,pwe,abc,pwe,xok,xyz,xxx,abc, roy,pwe,aaa,xxx,xyz,roy,xok.... how to send the unique data to one source and remaining data to another source????

3 Answers  


where we use config file as parameter ?

1 Answers   IBM,


Categories