I have source file which contains duplicate data,my
requirement is unique data should pass to one file and
duplicate data should pass another file how?
Answers were Sorted based on User's Feedback
Answer / dilip anand k
Its Simple!!
All you have to do is link your source to a Sort Stage.
Sort the data and generate a Key Change column.
Key Change column = ‘1’ represents that the record is
unique while Key Change Column = ‘0’ represents the
duplicates.
Put a Filter stage and filter out the data into two
different outputs based on the generated Key Change Column.
| Is This Answer Correct ? | 21 Yes | 5 No |
Answer / farzana kalluri
input output
1 T1 T2
2 4 1
2 6 2
1 7 3
3 4
4 5
3
5
5
6
7
for this
seq file---->Aggregate(key=id)---->filter---->2 targets
In aggregate use count rows...
in filter count=1 it goes to target1
if count=2 it goes to target2..
| Is This Answer Correct ? | 9 Yes | 3 No |
Answer / ramachandra rao
After source use aggregator stage and use option aggregator
type is count and count the records after that use filter in
where clause count>1 ie duplicate records go to one target
and another where clause count=1 ie unique records go to
another target.
| Is This Answer Correct ? | 3 Yes | 0 No |
Answer / sonali s
The above solution doesnt give required output. The requirement is as below:
Input:
A
B
B
C
D
D
D
Output should have 2 files as below.
File 1
A
C
File 2
B
B
D
D
D
Please provide solution for this
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / purba
Input:
A
B
B
C
D
D
D
Required output:
A
B
C
D
Solution:
Seq file----->sort stage(create key change column for the I/p key row)
O/p:
A 1
B 1
B 0
C 1
D 1
D 0
D 0
Now take filter stage to filter for key column=0 & keycol=1
We get 2 outputs:
A. B
B. D
C. D
D
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / riyazahamedmohamed
take two links using copystage, of your input file,one is your input file output, another one is for keychange column(using sort stage set the key change column to true) with filter "0" out of transformer, to the look up stage.set the lookup option to continue-reject.you will get the desired output.reject will capture unique records.output file will capture duplicate records.
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / krishna
As per my knowledge
initially soure is in sequential stage anc take aggrigator
stage and select the grouping option and select which column
you want to group then go to option command and select
column for calculation and select the which column you want
to do the operation .in column for calculation w have seen
many options and select missing count column name and give
the column name for output.and add transformer stage with in
the transformer stage add constraints .and give the two outputs
if column name=1 then 1 else 0
if column name>=2 then 1 else 0
it will work
| Is This Answer Correct ? | 0 Yes | 6 No |
I have a source table with column name CITY having 100 records, I want target table with column name start with 'A' and 'B',remaining columns as reject outputs. how can achieve this by data stage?please help me?????
If you want to use a same piece of code in different jobs, how will you achieve this?
what is data mapping
I have 100 records how can I load at a time from the single time
source has 2 fields like COMPANY LOCATION IBM HYD TCS BAN IBM CHE HCL HYD TCS CHE IBM BAN HCL BAN HCL CHE LIKE THIS....... AND I WILL GET THE OUTPUT LIKE THIS.... Company loc count TCS HYD 3 BAN CHE IBM HYD 3 BAN CHE HCL HYD 3 BAN CHE PLZ SEND ME ANSWER FOR THIS QUESTION..........
guys pls tell me where we use sequence jobs exactly in realtime proj explain pls with example.
I am having the 2 source files A and B and I want to get the output as, the data which is in file A and which doesn't in file B to a target 1 and which is in file B and which doesn't in file A to a target 2?
Two source files contains same meta data third file contains different data types can I funnel that file.
Wat is isolation level and when do u use them?
How can we read latest records in a text file named file1.txt using seq file stage only? file1 having 100 records in that 5 record sare latest records.How can we read that latest records?
Have you used Unstructured data?
what are .ctl(control files) files ? how the dataset stage have better performance by this files?