i/p o/p1 o/p2
1 1 4
1 1 5
1 1 6
2 2
2 2
2 2
3
3
4
5
6
how to populates i/p rows into o/p1&o/p2 using datastage
stages?and also the same scenario using sql?
Answers were Sorted based on User's Feedback
Answer / radhakant kumar
1)by sql say the column name is Id then
a)(for duplicate)
select a.id
from table_name a
left outer join(select id,count(*)as no1
from table_name
group by id
)b
on b.id=a.id
where b.no1>1
order by a.id
it will display 1 1 1 2 2 3 3
b)change the condition where b.no1=1 then it will display
the remainings without duplicates..
2) By DataStage
Step By Step....
a)after source file take Copy stage and draw two
output links
b)one will go to Aggregate stage and take count of
rows by group by Field(id)
it will display like.
ID Count(no)
1 3
2 3
3 2
4 1
5 1
6 1
c)and then use lookup stage main stream will be second link
of copy and reference will be output of Aggregate stage
d)look up key will be ID and fetch count(no) to target
it will display
ID Count(no)
1 3
1 3
1 3
2 3
2 3
2 3
3 2
3 2
4 1
5 1
e)and after lookup stage take transformer or filter stage
by taking two output links and put condition
a)for duplicates count(no)>1
b)for non duplicates count(no)=1
Finally u will get your desired output..
6
| Is This Answer Correct ? | 13 Yes | 1 No |
Answer / kalpana_241871
SQL>
SELECT COL1 FROM TAB1 GROUP BY COL1 HAVING COUNT(*) > 1 ;
O/P 1 :
1
1
1
2
2
2
3
3
SQL>
SELECT COL1 FROM TAB1 GROUP BY COL1 HAVING COUNT(*) = 1 ;
O/P 2 :
4
5
6
| Is This Answer Correct ? | 11 Yes | 2 No |
Answer / kavitha
sequentialfile-----copy-----head and tail(in that we have options like no of rows=6 for head and n..o..r=4 for tail)then connect two data sets to that head and tail u will get
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / rvanaja.etldeveloper@gmail.com
just read the data in sequential file
and take aggregator
group=cid
aggregation type=count rows
count output column=countno
then take filter from filter you drag 2 links for two outputs
in filter write conditions like bellow and give link order
where clause=countno>1
and
where clause=countno=1
map in output links
seqfile-->aggregator-->filter-->2 torgets
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / vijaya
Hi
first sort the input date then use the transformer stage by
using stage variables you can use the logic like check the
first record then check the second record if both are equal
send the output to o/p1 or else send to o/p2....
I hope i made u understand the logic
| Is This Answer Correct ? | 2 Yes | 11 No |
Answer / srinivas
first sort the data,
then
take two stage variable
if (sv1=sv2) then ds.link1 else ds.link2
| Is This Answer Correct ? | 1 Yes | 13 No |
What is the surrogate key? what is the use of surrogate key? how to Create surrogate key Generator in scd2 in 8.5?
1)i put Pharma Project in my Resume..whar are the sources used in my project Generally? 2)how many fact and dimensional tables used? 3)Have u used any Datamarts and measues in fact table? ....plz give the answers...
souce file having the columns like name company krish IBM pooja TCS nandini WIPRO krish IBM pooja TCS if first row will be repeat i want the result like this name company count krish IBM 1 pooja TCS 1 nandini WIPRO 1 krish IBM 2 pooja TCS 2
guys pls tell me where we use sequence jobs exactly in realtime proj explain pls with example.
Define data aggregation?
what is the difference between == and eq in UNIX shell scripting?
create a job that splits the data in the Jobs.txt file into four output files. You will direct the data to the different output files using constraints. • Job name: JobLevels • Source file: Jobs.txt • Target file 1: LowLevelJobs.txt − min_lvl between 0 and 25 inclusive. − Same column types and headings as Jobs.txt. − Include column names in the first line of the output file. − Job description column should be preceded by the string “Job Title:” and embedded within square brackets. For example, if the job description is “Designer”, the derived value is: “Job Title: [Designer]”. • Target file 2: MidLevelJobs.txt − min_lvl between 26 and 100 inclusive. − Same format and derivations as Target file 1. • Target file 3: HighLevelJobs.txt − min_lvl between 101 and 500 inclusive. − Same format and derivations as Target file 1. • Rejects file: JobRejects.txt − min_lvl is out of range, i.e., below 0 or above 500. − This file has only two columns: job_id and reject_desc. − reject_desc is a variable-length text field, maximum length 100. It should contain a string of the form: “Level out of range: <min_lvl>”, where <min_lvl> is the value in the min_lvl field. My Question is how do you write the stage variable for reject rows.
How many jobs in ur project? Explain any complex job u have done in ur project?
how to delete one row in target dataset
8000 jobs r there i given commit, suddenly job will abort? what happens? 2)diff b/t transformer stage & filter stage? 3)how to load the data in the source?
iam new to datastage...now i want to know what are fact tables, dimension tables in bank domain...if any body knows plz tell me asap..
Why do you need stage variables?