I have file with empid,empname and I want to load these two
fields along with sal in my target
1)salary must be same for all the records
2)I want pass the salary at run time
Answer Posted / shilpasagar
Declare one job parameter SAL_PARAM
Within the job seq -> Transformer -> Target
In Transformer, declare one new column by the name SALARY
and give its value as job parameter (select SAL_PARAM from
the list)
Compile the job and run it. While running it will prompt
you to pass value for SAL_PARAM.
If you pass SAL_PARAM as 45000, then all the emp data will
be populated with emp_id,empname,empsal (45000)
Note: Enable RCP too
Is This Answer Correct ? | 12 Yes | 0 No |
Post New Answer View All Answers
What a datastage macro?
What is a datastage job?
How to manage date conversion in Datastage?
Define Data Stage?
Is possible to create skid in dim,fact tables?
root tree will find which is server job and which is parallel job?
what is ds administrator used for?
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
What is the difference between an operational datastage and a data warehouse?
What are iconv and oconv?
Differentiate between operational datastage (ods) and data warehouse?
What is the differentiate between data file and descriptor file?
Can anyone tell me a difficult situation who have handled while creating Datastage jobs?
Can you explain link buffering?
Is it possible to query a hash file?