Create a job to load all unique products in one table and the duplicate rows in to another table.
The first table should contain the following output
A
D
The second target should contain the following output
B
B
B
C
C
Q2. Create a job to load each product once into one table and the remaining products which are duplicated into another table.
The first table should contain the following output
A
B
C
D
The second table should contain the following output
B
B
C
Answers were Sorted based on User's Feedback
Answer / sudheer
answer for Q2, use sort stage and use generate change key column option. use filter stage to send all change key columns having value=1 to move in to a file which generates A B C D, and filter records having value=0 to another file which generates B B C.
Is This Answer Correct ? | 7 Yes | 1 No |
Answer / unknown
Q1: First use Aggregate stage- Row Count property then filter stage to separate Row Count 1 and more than 1.
Is This Answer Correct ? | 4 Yes | 0 No |
Answer / nams
It can be done by unix uniq -u file_Name and uniq -d file_Name
or it can be done from datastage sequential file stage filter option from property.....write command overthere like above.....
Is This Answer Correct ? | 1 Yes | 3 No |
Define Routines and their types?
Define APT_CONFIG in Datastage?
Star Schema vs. Snow Flake Schema. What is the best model for DW? Give example for both the models performance wise?
hi everbody, i have a problem in my project.. Native Error Code: -407 "Assignment of a null value to a not null column" hi everbody, i have a problem in my project.. Native Error Code: -407 "Assignment of a null value to a not null column" i have no idea, what should i do..please help :(
Converting Vertical PIVOTing without using PIVOT stage in DataStage. Ex: DEPT_NO EMPNAME 10 Subhash 10 Suresh 10 sravs Output: DEPT_NO EMP1 EMP2 EMP3 10 subhash suresh sravs 2) How to implement Horizontal PIVOTing without using PIVOT stage.
create a job that splits the data in the Jobs.txt file into four output files. You will direct the data to the different output files using constraints. • Job name: JobLevels • Source file: Jobs.txt • Target file 1: LowLevelJobs.txt − min_lvl between 0 and 25 inclusive. − Same column types and headings as Jobs.txt. − Include column names in the first line of the output file. − Job description column should be preceded by the string “Job Title:” and embedded within square brackets. For example, if the job description is “Designer”, the derived value is: “Job Title: [Designer]”. • Target file 2: MidLevelJobs.txt − min_lvl between 26 and 100 inclusive. − Same format and derivations as Target file 1. • Target file 3: HighLevelJobs.txt − min_lvl between 101 and 500 inclusive. − Same format and derivations as Target file 1. • Rejects file: JobRejects.txt − min_lvl is out of range, i.e., below 0 or above 500. − This file has only two columns: job_id and reject_desc. − reject_desc is a variable-length text field, maximum length 100. It should contain a string of the form: “Level out of range: <min_lvl>”, where <min_lvl> is the value in the min_lvl field. My Question is how do you write the stage variable for reject rows.
how to connect source to db?generally what r stages u used? how to find the data is having delimiter format?
what is the Difference Between Datastage Server Edition and Parallel Edition?
iam new to datastage...now i want to know what are fact tables, dimension tables in bank domain...if any body knows plz tell me asap..
how may datastage variables/parameters will be in trnsformer stage? 1 2 3 4 ?
1)How to do error handling in datastage? 2)Did sequential stage accepts .xl files ,xml? znd how?
Can you explain repository tables in datastage?