I am running a job with 1000 records.. If the job gots
aborted after loading 400 records into target... In this
case i want to load the records in the target with 401
record... How will we do it??? This scenario is not for
sequence job it's only in the job Ex: Seq file--> Trans-->
Dataset..
Answer Posted / phani kumar
With the help of available Environment variable as
APT_CHECKPOINT_DIR, we can run the job for remaining records.
This is available in Datastage Administrator like.....
With Datastage Administrator - project-wide defaults for
general environment variables, set per project in the
Projects tab under Properties -> General Tab -> Environment
variables.
Here, we can enable the this checkpoint variable. Then, we
can load the remaining records........
| Is This Answer Correct ? | 6 Yes | 3 No |
Post New Answer View All Answers
What is developer responsibilities in UAT (user acceptance testing and Post implementation phase?
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
What are stage variables and constants?
Differentiate between Symmetric Multiprocessing and Massive Parallel Processing?
State the difference between an operational datastage and a data warehouse?
What is the use of datastage director?
What is the process of killing a job in datastage?
Can you highlight the main features of ibm infosphere information server?
What is a ds designer?
Where do the datastage jobs get stored?
How the ipc stage work?
disign the complex job in u r project?(they are aksing only complex job design and then data flow...)
Explaine the implimentation of scd's in ds indetail, please send me step by step procedure to perform scd's 1,2,3. Please replay for this, Thanks in advance
What are the components of datastage?
What is the use of hoursfromtime() function in transformer stage in datastage?