Create a job to load all unique products in one table and the duplicate rows in to another table.
The first table should contain the following output
A
D
The second target should contain the following output
B
B
B
C
C
Q2. Create a job to load each product once into one table and the remaining products which are duplicated into another table.
The first table should contain the following output
A
B
C
D
The second table should contain the following output
B
B
C
Answers were Sorted based on User's Feedback
Answer / sudheer
answer for Q2, use sort stage and use generate change key column option. use filter stage to send all change key columns having value=1 to move in to a file which generates A B C D, and filter records having value=0 to another file which generates B B C.
| Is This Answer Correct ? | 7 Yes | 1 No |
Answer / unknown
Q1: First use Aggregate stage- Row Count property then filter stage to separate Row Count 1 and more than 1.
| Is This Answer Correct ? | 4 Yes | 0 No |
Answer / nams
It can be done by unix uniq -u file_Name and uniq -d file_Name
or it can be done from datastage sequential file stage filter option from property.....write command overthere like above.....
| Is This Answer Correct ? | 1 Yes | 3 No |
SEQUENTIAL FILE I HAVE ONE RECORD,I WANT 100 RECORDS IN TARGET?HOW CAN WE DO THAT?PLS EXPLAIN ME AND WHAT STAGES ARE THERE?WHAT LOGIC?
CAN WE DECLARE DATASET SET WITH OTHER THAN .DS EXTENTION OR NOT IF YES WHAT IS THE DIFFERENCE U FIND?IF NOT WHY?
what is the difference between lookup stage reject link and merge stage reject link in datastage Parallel jobs?
How to enter a log in auditing table whenever a job get finished?
What is the difference between validated and compiled in the datastage?
DB2 connector> transformer > sequential file Data will be exported into a csv format in a sequential file. This file will be send in a email using a sequence job. Problem here is, how to avoid sending a blank csv file? When I ran the job there are chances that it might return zero records but in the sequence job csv file is going blank. how can I avoid this? thanks
A flat file contains 200 records. I want to load first 50 records at first time running the job, second 50 records at second time running and so on, how u can develop this job?
Tell me the syntax of Configuration file?
Converting Vertical PIVOTing without using PIVOT stage in DataStage. Ex: DEPT_NO EMPNAME 10 Subhash 10 Suresh 10 sravs Output: DEPT_NO EMP1 EMP2 EMP3 10 subhash suresh sravs 2) How to implement Horizontal PIVOTing without using PIVOT stage.
my source is sequencial file and my target is dataset. i am running the job in two node configuration file. my source having 10 records how the data move to target?
How one source columns or rows to be loaded in to two different tables?
Differentiate between datastage and informatica?