Instead of using shared container in a job, I use jobs
which perform similar function as Container in the sequence.
Then what is the need of Shared Container?
Answer Posted / vishwanath
You can make the shared container as Local or Global.
If declared as global, we can use the shared container across the jobs. This just require one time compilation. If the job is used instead of Shared Container, it will be compiled where ever the job is used.If the job is big containeing transformations, just imagime the performance..... hope i'm clear with the limited knowledge I have :)
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers
How do you remove duplicate values in datastage?
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
Hi , Can anyone give few examples of scenarios and there corresponding design in datastage..i am new to this tool...confused in design while my manager asking to design the job.. Please post the URL if there..so i can go through it.. Thanks in advance...
Enlist various types of routines in datastage.
Describe link sort?
1)s.key generate 1 to 700 records today. tomorrow another 400 will updated how to update the records using s.key generator? 2)source is like :-- DB --> T/F stage1 --> seq1file T/f 1 is linking with T/F2 ---> seq 2 how to load the data? in source i given some conditions those r going in seq1. The another data will going to seq2 how to do this ?
What are the primary usages of datastage tool?
Which warehouse using in your datawarehouse
What is "fatal error/rdbms code 3996" error?
How do you schedule or monitoring the job?
Could anyone give brief explanation bout datastage admin
What are the types of jobs we have in datastage?
If we take 2 tables(like emp and dept),we use join stage and how to improve the performance?
What are the types of hashed files in data stage
How to implement complex jobs in data stage?