Source flat file contains
src
---
1
2
'
'
'
18
we had 3 targets
T1 T2 T3
-- -- --
1 4 7
2 5 8
3 6 9
10 13 16
11 14 17
12 15 18
How can i get?
**Using only datastage, but not unix or any other.
I am expecting the answer soon..
Thanks in advance.
Answer Posted / pranay
source----->transformer----->filter------>T1
|
|
--------->T2
|
|
--------->T3
Transformer:
1.drag source column to target.
2.append new column in transformer target as new and apply derivation Mod(input column, 3).
Filter:
where condition1 new = 1
condition2 new = 2
condition3 new = 0
Name the three targets T1, T2, T3.
| Is This Answer Correct ? | 3 Yes | 3 No |
Post New Answer View All Answers
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
what are the devoleper roles in real time? plz tell i am new to datastage....
How and where you used hash file?
What is the process of killing a job in datastage?
Where the datastage stored his repository?
What are the steps required to kill the job in Datastage?
How to reverse the string using unix?
Can you explain kafka connector?
1.new record it will insert but changes of natural key is not present in taget i want to update (here key is composite natural key )can any one help this to explan how to do
In Datastage, how you can fix the truncated data error?
i have a job with 3 possibiliies finish with ststus ok abbort warning how do you handle these 3 in job sequence? explain the situation where you have appiled scd in your project? chandu
What is size of a transaction and an array means in a datastage?
how to export or import the jobs in .ISX file
How to implement complex jobs in data stage?
How will you move hashed file from one location to another location?