if the source file is
CID,CCODE,CONNDATE,CREATEDBY
0000000224,1000,20060601,CURA
0000000224,2000,20050517,AFGA
0000000224,3000,20080601,TUNE
0000000225,1000,20020601,CURA
0000000225,2000,20050617,AFGA
0000000225,3000,20080601,TONE
AND TARGET is oracle
following are the validations
cid loaded with unique records
leading zeors has to be deleted while loading cid in target
load only customer who got early connected to company
conn_date should be loaded into oracle date format
cid datatype is varchar2 in target
conn_date is data datatype
ccode is varchar2
0000000224,1000,20060601,CURA
0000000224,1000,20060601,CURA
Answer Posted / supriya
Use Sort stage to get unique recs,write a constraint in the
transformer to remove leading zeros by usig Trim
function.In XFM it sims we can change the datatypes of
required fields.
| Is This Answer Correct ? | 0 Yes | 2 No |
Post New Answer View All Answers
Enlist various types of routines in datastage.
Hi,can any one please mention list of dimension and fact tables for Sales and Distribution System for a Pharmacutical company.
Where the datastage stored his repository?
What are the main features of datastage?
What is the use of datastage designer?
What is the difference between orabulk and bcp stages?
Have you have ever worked in unix environment and why it is useful in datastage?
Name the different sorting methods in datastage.
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
if we using two sources having same meta data and how to check the data in two sources is same or not? and if the data is not same i want to abort the job ?how we can do this?
if i have two tables table1 table2 1a 1a,b,c,d 1b 2a,b,c,d,e 1c 1d 2a 2b 2c 2d 2e how can i get data as same as in tables? how can i implement scd typ1 and type2 in both server and in parallel? field1 field2 field3 suresh , 10,324 , 355 , 1234 ram , 23,456 , 450 , 456 balu ,40,346,23 , 275, 5678 how to remove the duplicate rows,inthe fields?
What are constraints and derivations?
What are iconv and oconv?
What are the types of containers and how to create them?
DB2 connector> transformer > sequential file Data will be exported into a csv format in a sequential file. This file will be send in a email using a sequence job. Problem here is, how to avoid sending a blank csv file? When I ran the job there are chances that it might return zero records but in the sequence job csv file is going blank. how can I avoid this? thanks