Hi guys,

In sequencer job,
we have 3 sources, in that 1st source have some records,
Here requirement is 1st source records are 100 then only run
the job otherwise total job will abort...
How to calculate this. please design the job.

Thanks.

Answer Posted / anil

Hi,

We can do it with one seq job and one parallel job

First create parallel job (p1) that calculates aggregate count. If aggregate count is less than 100 then abort the parallel job

Paralled Job Design(P1)

1. Take the data from source.
2. Add a column generator with column name as Temp and its value as A
3. Add aggregrator stage, where group by column is Temp and calculation is count rows. So after aggregation, count will be calculated and stored in Total Count column.
4. In the transformer, add a constraint for output link to a peak stage saying totalcount < 100 and abort after 1 row.

SO if count is greater than 100, parallel job (P1) is successful, Else failed

Sequence Job

1. Add parallel job(p1) to the sequence and add exception handler and terminator activity stage.

2. After P1 you can create whatever sources you want to as a each parallel job.

3. Add automatically handle failed activities in the sequence job properties

Now if P1 fails, sequence fails else rest will run.

Is This Answer Correct ?    2 Yes 0 No



Post New Answer       View All Answers


Please Help Members By Posting Answers For Below Questions

How to write a expression to display the first letter in Caps in each word using transformer stage ? Please let me know ASAP Thanks in advance...

1080


How do you remove duplicate values in datastage?

630


what is repositery?

1297


What are orabulk and bcp stages?

649


how to add a new records into source?

1529






Define project in datastage?

666


What are the primary usages of datastage tool?

623


What are the different kinds of views available in a datastage director?

603


Hi,can any one please mention list of dimension and fact tables for Sales and Distribution System for a Pharmacutical company.

1669


create a job that splits the data in the Jobs.txt file into four output files. You will direct the data to the different output files using constraints. • Job name: JobLevels • Source file: Jobs.txt • Target file 1: LowLevelJobs.txt − min_lvl between 0 and 25 inclusive. − Same column types and headings as Jobs.txt. − Include column names in the first line of the output file. − Job description column should be preceded by the string “Job Title:” and embedded within square brackets. For example, if the job description is “Designer”, the derived value is: “Job Title: [Designer]”. • Target file 2: MidLevelJobs.txt − min_lvl between 26 and 100 inclusive. − Same format and derivations as Target file 1. • Target file 3: HighLevelJobs.txt − min_lvl between 101 and 500 inclusive. − Same format and derivations as Target file 1. • Rejects file: JobRejects.txt − min_lvl is out of range, i.e., below 0 or above 500. − This file has only two columns: job_id and reject_desc. − reject_desc is a variable-length text field, maximum length 100. It should contain a string of the form: “Level out of range: ”, where is the value in the min_lvl field. My Question is how do you write the stage variable for reject rows.

2195


What is the difference between orabulk and bcp stages?

576


what is flow of project?

1559


Why we use surrogate key?

699


How to read the length of word in unix?

885


What is the method of removing duplicates, without the remove duplicate stage?

586